Table of Contents Author Guidelines Submit a Manuscript
Computational Intelligence and Neuroscience
Volume 2019, Article ID 3761203, 11 pages
https://doi.org/10.1155/2019/3761203
Research Article

A LightGBM-Based EEG Analysis Method for Driver Mental States Classification

1School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, Zhejiang 310018, China
2Industrial NeuroScience Lab, University of Rome “La Sapienza”, Rome 00161, Italy

Correspondence should be addressed to Wanzeng Kong; nc.ude.udh@gneznawgnok

Received 6 February 2019; Revised 29 June 2019; Accepted 22 July 2019; Published 9 September 2019

Guest Editor: Lewis Chuang

Copyright © 2019 Hong Zeng et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Fatigue driving can easily lead to road traffic accidents and bring great harm to individuals and families. Recently, electroencephalography- (EEG-) based physiological and brain activities for fatigue detection have been increasingly investigated. However, how to find an effective method or model to timely and efficiently detect the mental states of drivers still remains a challenge. In this paper, we combine common spatial pattern (CSP) and propose a light-weighted classifier, LightFD, which is based on gradient boosting framework for EEG mental states identification. The comparable results with traditional classifiers, such as support vector machine (SVM), convolutional neural network (CNN), gated recurrent unit (GRU), and large margin nearest neighbor (LMNN), show that the proposed model could achieve better classification performance, as well as the decision efficiency. Furthermore, we also test and validate that LightFD has better transfer learning performance in EEG classification of driver mental states. In summary, our proposed LightFD classifier has better performance in real-time EEG mental state prediction, and it is expected to have broad application prospects in practical brain-computer interaction (BCI).

1. Introduction

Fatigue driving is an important cause of traffic accidents. According to data from U. S. National Transportation Safety Board, the annual economic losses caused by driving accidents in the United States are more than $12.5 billion [1]. Fatigue has no obvious symptoms but usually manifests as lethargy, fatigue, or weakness [2]. Therefore, developing technologies to monitor and predict driver’ mental state or the ability to safely drive the vehicle will have significant social and economic benefits [3].

At present, for fatigue driving detection, the academic community has carried out a lot of research work. To sum up, it mainly lies in the following aspects: (1) mental activity testing using response time and accuracy by passive BCIs [4, 5], which mainly perform an assessment of a subject’s cognitive states [6, 7], (2) detection of eye movement parameters, such as eye squint movement, percentage closure of eyes (PERCLOS) [8], and so on, (3) active detection by means of questionnaires, (4) sensor-based methods to find some fatigue indicators by steering force (steering grip pressure), skin conductance, blood volume pulse (BVP), and so on [9, 10], and (5) performing fatigue state detection by bioelectrical signals, such as EEG, EOG (electrooculogram), EMG (electromyogram), and ECG (electrocardiogram) [1116].

For physiological-electric-based detection, researches have shown that these signals have a strong correlation with the driver’s mental state, so these signals can be more accurate to detect driving fatigue. Among the above various researches of fatigue detection, EEG analysis methods are considered to be most convenient and effective for its good time resolution and sufficient spatial resolution. It is known that EEG represents the brain activity by the electrical voltage fluctuations along the scalp [17]. As an effective tool for the indirect measurement of neural activity, EEG is widely used in neuroscience, cognitive science, cognitive psychology, and psychophysiology research, etc. On the other hand, driving behavior involves a variety of behaviors, such as motions, reasoning, audiovisual processing, decision making, perception, and recognition, which is also affected by emotions, attention [18], and many other psychological factors. These physical and mental activities related to driving are reflected in EEG signals.

In recent years, a number of methods for fatigue detection using EEG have been proposed; for example, Kar et al. [2] investigated a number of fatigue-indicating parameters based on higher-order entropy measures of EEG signals in the wavelet domain. In particular, they present a method based on a kind of entropy measures on the EEG signals of the subjects for the relative quantification of fatigue during driving. Charbonnier et al. [19] proposed an online innovative EEG index and proved that the proposed index can be used based on the alpha activity to effectively assess the operator’s mental fatigue status. Roy et al. [20] applied a Fisher’s linear discriminant analysis (FLDA) to detect and classify EEG-based mental fatigue. In [21], the authors used a KPCA-SVM classifier to distinguish between normal and fatigue mental state, with an accuracy rate of 98.7%. Maglione et al. [22] used high-resolution EEG and neurophysiological variables to analyze the increase in cerebral workload and the insurgence of drowsiness during car driving and acquired a workload index. In 2014, Zhang et al. [23]presented a real-time method with various entropy and complexity measures for the detection and identification of driving fatigue from EEG, EMG, and EOG signals, and the accuracy of estimation is about 96.5%–99.5%. Appriou et al. [24] presented a comparison of 4 modern machine learning algorithms in order to compare EEG-based workload level classification performances and found that CNN can obtain better performance (mean = 72.7% ± 9.1) than a LDA classifier with CSP spatial filters in classifying two workload levels (low vs. high) for both user-specific and user-independent studies. In [25], the authors developed an adaptive stacked denoising auto encoder (SDAE) to tackle cross-session mental workload (MW) classification task, and the adaptive SDAE is also demonstrated to be acceptable for online implementation. All together, these articles support the knowledge that mental fatigue can be efficiently detected by EEG with classification performances varying between 75% and 98%.

Other feature extraction and analysis methods are also used in mental state detection, such as EEG and fNIRS joint analysis [26], discrete wavelet transform [27], wavelet-packets transform (WPT) [28], integrating feature selection, and fusion on high-level EEG features from different models [29]. In recent years, deep learning-based models have also been used in mental state classification, for instance, deep convolutional neural networks [29], long short-term memory network (LSTM) [30], and switching deep belief networks with adaptive weights (SDBN) [31].

Although these methods have achieved excellent performance, how to design appropriate models to obtain robust, real-time, and high-accuracy classification performance of driving mental states by EEG still remains a challenge for a series of reasons. First, EEG shows the characteristics of instability and randomness, EEG signals collected by the single subject (intrasubject) or between two different subjects (intersubject) tend to have large differences over time [32]. Second, the low signal-to-noise (SNR) ratio of EEG often affects the accuracy of detection. Third, with the continuous improvement in EEG acquisition equipment, EEG signals gradually show multidimensional and complex features with a large time and space consumption during processing.

LightGBM [33] is a gradient boosting framework that uses a decision tree-based learning algorithms. It is distributed, efficient with faster training efficiency, and can handle a large amount of applications, but there also exists deficiencies when dealing with high-dimensional features for EEG signals, like lower accuracy, as well as time consumption. Therefore, in this article, we improve and design a LightGBM-based model, LightFD, which adopts the histogram-based decision tree algorithm and the leafwise leaf growth strategy with depth limitation to solve the problem of excessive xgboost memory consumption, which is more suitable for practical EEG clinical applications. Now, LightGBM has been applied to EEG signal classification and has achieved certain results in practical problems, such as emotion recognition [34, 35], epilepsy prediction [36], and so on.

Transfer learning methods have been widely used for EEG signal classification in recent years [3740], which could transfer the previous extracting features in one kind of trained samples to another sample for some specific decision tasks. Due to its great advantages of lower time consumption, transfer learning can bring more practical application possibilities for EEG analysis.

Motivated by the advantages of LightGBM and following our previous work [41], where only a CNN-based model was investigated to realize the EEG-based binary classification of mental states, and the model is time consuming, moreover, the transfer learning capability of the model is not analyzed. Thus, in this article, we aim to design a LightGBM-based classifier, LightFD, to implement the light-weighted analysis of triclassification identification of EEG mental states, and furthermore, we will also test and validate the efficiency and robustness of LightFD in the aspect of transfer learning and compare them with those of manifold embedded distribution alignment (MEDA) [42] and metric transfer learning (MTLF) [43].

2. Materials

2.1. Subjects

We recruited 10 healthy subjects for EEG data collection. All of them were within 23 and 25 years old and possess Chinese manual driver C1 license. They were informed in advance of the entire experimental process and instructions and also required to keep calm without drinking irritating beverage, like coffee, alcohol, and so on before the experiment. All participants provided their written consents, and the research was approved by the ethics committee of our university.

2.2. Experimental Setup

To collect EEG data during driving, we constructed a simulation platform, as shown in Figure 1, which consisted of a racing seat cushion, steering wheel, liquid crystal display (LCD), speaker, video camera, and projector. A 16-channel gUSBamp amplifier (g.Tec Medical Engineering GmbH) was used to record EEG signal. Besides, two more computers were employed for (1) simulating the track with the special “Speed-Shift 2 Unleashed (NFS-S2U)” software, recording all the parameters during driving with “WorldRecord” software, and (2) collecting the video and sound stimuli, dealing with EEG signals, respectively [12, 41].

Figure 1: Driving simulation experiment platform.
2.3. Experimental Protocol

The whole experiments lasted for two days and were conducted between 18 : 00 and 21 : 00 in a quiet and isolated environment. The first day was considered the practice stage for familiarizing with the track and stimulating software and experimental operations, and the second day was the formal experimental stage for collecting EEG data. The heart rate and blink were simultaneously collected with EEG by corresponding sensors, such as ECG electrode attached on the subject’s wrist and video camera placed in front of the subjects, which were used to aid judgment in the level of mental states. According to previous studies [44, 45], the numbers of blink and heart rate in the situation of awake would be higher than those in the situation of drowsiness. Moreover, we counted the average number of blink and heart rate of all the subjects throughout the whole experiments. We found that at the beginning stages, when the subjects were only asked to drive a car at a predefined speed without any video or sound stimuli, the average number of blinks is above 20 times/min, up to 24 times/min, and the average heart rate was close to 90 times/min. As the stimuli were introduced into the experiments, the average number of blinks changed to 12–20 times/min, and the heart rate was 78–85 time/min. At the last stage, the subjects were not given any stimuli, the average number of blinks increased to 22 times/min, and the average heart rate was 73 times/min. Therefore, we divided the mental states into 8 stages: WUP, PERFO, TAV3, TAV1, TAV5, TAV2, TAV4, and DROWS [12, 41]; the detailed introduction of these eight stages is shown in Table 1.

Table 1: Eight experimental stages.

The flowchart of the experiments is shown in Figure 2. There were two kinds of driving tasks during the experiments: one was a simple driving task, which only required the subject to drive like the practice stage and did not exert any sound and video stimuli. This kind of driving tasks included three stages: WUP, PERFO, and DROWS. WUP was the beginning of the experiment with a baseline driving speed, and PERFO was similar to WUP but required the subjects to drive at a speed of 2% faster than WUP. DROWS was the last stage of the experiments with a fixed driving speed of 60 km/h. The other stages introduced additional video (“alert”) and sound stimuli (“vigilance”) to simulate situations such as red lights and traffic jams that might occur in real driving, which include five TAV stages: TAV1–5. All TAV stages were exerted with different stimulus frequencies of sound (“vigilance”) and video (“alert”) stimuli that appeared on the LCD screen 1 m ahead of the subjects, and the corresponding buttons were pressed by the subjects: LEFT button for “vigilance,” and RIGHT button for “alert.” Five TAV stages: TAV3, TAV5, TAV1, TAV2, and TAV4 were executed in sequence.

Figure 2: The schematic diagram of experiment procedure.

Because TAV3 is the first stage with video and sound stimuli, the subjects were bound to drive very carefully and complete the corresponding operations as quickly and accurately as possible, so they were in the most awake state. DROWS was the last stage without any stimuli. It just required the subjects drive with a fixed speed of 60 km/h. It seemed monotonous and boring, especially after about 2 h of driving; therefore, the subjects were extremely prone to fatigue in this stage. In addition, the obvious differences in blink and heart rate between TAV3 and DROWS further confirmed the correctness of the design of this experiment. Moreover, we defined a “neutral” stage as TAVX, which was neither fatigue nor awake. However, the time required for each subject to enter the fatigue state may be different, and TAVX is one of those 4 stages: TAV1, TAV2, TAV4, and TAV5, at which the number of rushing out of the track is closest to the average of rushing out of the track during the experiment. Accordingly, the collected data at TAVX were then used for analysis.

2.4. EEG Recording

EEG was recorded by a gUSBamp amplifier with a sampling frequency of 256 Hz and impedance of below . Of 16 channel electrodes, 15 were used to sample EEG, except for ECG sampling heart rate. All the electrodes were referenced to the left earlobe. After removing the artifacts, EEG signals of 15 channels were divided as Fz, Pz, Oz, Fp1, Fp2, F7, F3, F4, F8, C3, C4, P7, P3, P4, and P8, with a time window of 0.5 s, and then a certain number of epochs of each stage were obtained. According to Kar et al. [2], the EEG recording was filtered between 1 and 40 Hz with a band-pass filter, and independent component analysis (ICA) [46] was then adopted for eye movement artifacts rejection. With ICA, the source signal can be separated or approximately separated without knowing the source signalS, noise, and mixing mechanism. After that, according to the method we proposed in [41], EEG recording was converted into SP  CH  TR format, where SP is the sampling frequency, CH is the corresponding channel, and TR is the event. For the segmentation of EEG data, we adopted 0.5 s-interval time window to split EEG data of 15 channels into different number of epochs. Due to the sampling frequency of 256 Hz, we then expressed every epoch as a 15  128 matrix. At the same time, we used the flag “0” for DROWS, “1” for TAV3, and “2” for TAVX, respectively. Thus, we obtained a total of 37,168 epochs, including 18,672 DOWNS epochs, 9,504 TAV3 epochs, and 8,992 TAVX epochs, as shown in Figure 3. In this way, LightFD can be trained by these epochs, and the classification performance of LightFD for mental states prediction could be tested simultaneously as well.

Figure 3: Epoch number of DROWS, TAV3, and TAVX for the subjects.

3. Method

3.1. EEG Feature Extraction by Improved CSP

The core of CSP is to find the optimal spatial projection to maximize the power of the two types of signals, so it can estimate two spatial filters to extract the task-related signal components and remove the task-independent components and noise. The method used by CSP is based on the simultaneous diagonalization of two covariance matrices.

For EEG data we extracted, each trail can be represented as a matrix W of , where X is the number of channels and S is the number of sampling points for each channel. The regularized spatial covariance is shown in the following equation:where trace() represents the sum of the diagonal elements of the matrix. In order to separate the two types of variances, we averaged the sum of the covariances of the two types of samples in the training data to achieve the respective average covariances and and then obtained the mixed spatial covariance as  = . , which was decomposed into the form  = , where is the eigenvector of the matrix and is the diagonal matrix formed by the eigenvalues. The eigenvalues were arranged in descending order, and the whitening transformation was performed according to the following equation:

The eigenvalue corresponding to is 1, so and were transformed as follows: , . Then, and share common feature vectors; when , there are and , where I is the unit vector matrix. Because the sum of the corresponding two eigenvalues is always 1, when eigenvector B has the largest eigenvalue for , it has the smallest eigenvalue for . Thus, the projection matrix obtained was

Because three states were used in the experiment, we designed a feature extraction method for the three categories by CSP. For the awake state that was easier to distinguish; we projected the fatigue state data and the neutral state data separately and obtained the projection matrices and . Our final projection matrix was

All experimental samples (including training and testing) were decomposed according to equation (4) to obtain the required EEG characteristics:

The process of EEG feature extraction is shown in Figure 4. In addition, the high dimensionality of EEG data increased the time and space consumption in deep learning models. But through our experimental tests, we found that LightGBM did not rely on high-dimensional data features as deep learning models. After feature reduction, the training speed was faster, memory consumption was reduced, and the final accuracy did not change much.

Figure 4: EEG data feature extraction process.

Traditional CSP usually uses log variance for feature normalization in the binary-classification problem. While in our proposed improved CSP for EEG-based triclassification problem, after obtaining the feature matrix through the projection matrix W, instead of using the conventional method, we used the channel variance of the feature matrix to achieve the purpose of dimensionality reduction. At last, the variance function: var, for each sample, was used to calculate the variance of the data in each channel and reduced the dimensionality of EEG data. Based on this improved CSP, we designed and implemented a LightGBM-based model, LightFD, for the triclassification of driver mental states.

3.2. Training of LightFD Classifier

LightGBM is an algorithm for classification that relies on the gradient hoist, and it is known for its light computational burden [33]. In particular, in the tree-based boosting family of algorithms, many of them (such as xgboost) use the presorting algorithm to select and split features. However, this presorting algorithm can accurately find the splitting point, but it has a large overhead in time and memory consumption. The proposed LightFD model adopts histogram algorithm and leaf growth strategy of leafwise with depth limitation, as shown in Figure 5, which can increase computing efficiency, decrease memory occupancy, improve classification accuracy, and prevent overfitting efficiently (please refer to [33, 47] for more detail). The detailed procedure of LightFD is listed as follows.

Figure 5: Learning process of LightGBM.
3.2.1. Histogram Algorithm

The basic idea of the histogram algorithm is to discretize successive floating-point eigenvalues into k integers and construct a histogram of width k. When traversing the data, the statistic is accumulated in the histogram according to the discretized value as an index. After traversing the data once, the histogram accumulates the required statistic and then traverses to find the optimal segmentation point according to the discrete value of the histogram.

3.2.2. Leafwise Leaf Growth Strategy with Depth Limitation

Levelwise data can split the leaves of the same layer at the same time, easy to multithread optimization, control model complexity. But levelwise is actually an inefficient algorithm because it treats the leaves of the same layer indiscriminately, which brings a lot of unnecessary overhead, and is difficult to prevent overfitting, due to the lower split gain of many leaves, which does not need to be searched and split.

Leafwise strategy is more efficient. It is just to find the leaf that has the highest split gain from the current layer to split. Therefore, compared with levelwise method, leafwise strategy can obtain better performance at the situation with the same number of split. But leafwise strategy may cause deeper decision tree and then be overfitting. So to avoid the situation of overfitting and ensure higher efficiency, we then make a maximum depth limitation in LightFD model.

3.3. Parameters of LightFD

The parameters in lightFD include num_leaves, num_trees, and learning_rate, where num_trees represents the total number of spanning trees and num_leaves represents the number of leaves on per spanning tree. Smaller learning_rate and larger num_trees can improve the final accuracy to a certain extent, but it increases the time and space overhead.

4. Results and Discussion

As traditional machine learning methods, SVM [48] and LMNN [49] are classical methods for the classification of samples. Deep learning (DL) [50] has been successfully applied in many fields such as computer vision, speech recognition, and natural language processing. LSTM is proposed to overcome the fact that the recurrent neural network (RNN) does not handle long-range dependencies well, although GRU is a variant of LSTM. GRU maintains the effects of LSTM with a simpler structure and plays its own advantages in more and more fields. CNN is a neural network designed to process data similar to grid structures, such as time series data and image data, which has become one of the most important representatives of DL because of its excellent classification performance in many challenging applications [5153].

In this section, we compare LightFD with SVM, LMNN, GRU, and CNN from both aspects of intrasubject and intersubject. Particularly, because GRU and CNN rely on high-dimensional features, we do not perform dimensionality reduction after CSP but directly use those high-dimensional features as input to GRU and CNN models for training and testing.

For SVM, the kernel type is used with a Gaussian kernel function, the penalty parameter is set to 1.5, and probability estimation is set as “not enabled.” For LMNN, we chose the Euclidean distance as the distance metric, and the nearest neighbor is set to 3. For GRU, we used a single-layer structure with a time step of 128, a learning rate of 0.001, and the RMSprop model as the gradient descent method. For CNN, the model structure contains a convolutional layer (output number is 32) and a convolutional layer (output number is 32 as well), followed by a maximum pooling layer with a step size of 2, and a learning rate of 0.01.

4.1. Intrasubject Classification Performance

For each subject, we randomly extracted 80% of EEG signals as a training set, denoted as Train_i, and the remaining 20% as a test set, denoted as Test_i, where i = 1, 2, … , 10, indicating the i-th subject, the ratio of the training set to the test set is strictly 4 : 1; both Train_i and Test_i were the data sets after dimensional reduction by our improved CSP.

When comparing LightFD with SVM and LMNN, we adopted Train_i and Test_i as the training set and test set, respectively, for the analysis of classification performance. Although comparing LightFD with GRU and CNN, due to the high-dimensional feature correlation of GRU and CNN, we did not adopt those features processed by the improved CSP as input but the original data after preprocessing.

The test results are shown in Figure 6. For the mental state detection of the same subject (intrasubject), SVM and LMNN models have similar classification performance, and their average classification accuracy are 90.10% and 88.10%, respectively; however, LightFD reaches the average accuracy of 95.31%, which is much higher than others. GRU and CNN only surpass LightFD in the classification performance of subject s4, whereas the classification accuracy of the other 9 subjects is inferior to that of LightFD.

Figure 6: Accuracy comparison of SVM, LMNN, CNN, GRU, and lightFD for intrasubject classification.

In addition, we also counted the average classification accuracy of those 5 models for intrasubject, as shown in Table 2. We found that LightFD has the best classification performance among these models.

Table 2: Average classification accuracy of SVM, LMNN, GRU, CNN, and LightFD for intrasubject.

To evaluate the stability performance of LightFD, we then calculated the variance of the accuracy of LightFD, SVM, LMNN, GRU, and CNN, respectively, as shown in Table 3.

Table 3: Variance analysis of SVM, LMNN, GRU, CNN, and LightFD for intrasubject.

From Table 3, it is clear that SVM and LMNN have some extent of similar stability and better than GRU and CNN, but the variance of LightFD is significantly lower than those of all others, which shows that LightFD has better robustness in EEG signal processing and further lays the foundation for its real application.

Moreover, to validate the applicability of LightFD, we randomly divided the existing data sets for 5 times and obtained 5 groups of data containing different test sets and training sets, then tested the performance of LightFD by the 5 groups of training and test sets. The acquired results are shown in Figure 7. From Figure 7, it is clear that the different data set has a certain impact on the classification result, for example, the average accuracy of subject s4 decreases to about 88%, but in a whole, LightFD keeps a much higher classification accuracy under different test data sets.

Figure 7: Classification accuracy statistics of 10 subjects under the condition of different testing sets.
4.2. Intersubject Classification Performance

EEG signals vary widely among subjects, and these differences can affect the final classification results. To further test the performance of LightFD, in this section, we made a classification performance analysis of intersubject.

Similarly, we mixed all the EEG data of 10 subjects and randomly selected 80% of them as training sets, the remaining 20% as test sets. We also conducted the classification performance analysis and comparison for intersubject analysis between SVM, LMNN, GRU, CNN, and LightFD. To satisfy the input need of GRU and CNN, we do not yet carry out the operation of dimensionality reduction for the two models.

As shown in Figure 8, LightFD has a classification accuracy of 91.67%, which is significantly higher than SVM with 74.54%, LMNN with 57.59%, GRU with 73.19%, and CNN with 77.89%. The comprehensive performance of CNN for intersubject analysis is slightly better than SVM but much lower than LightFD. Also from the intersubject classification results, it was found that, compared with intrasubject test, LightFD could maintain more stable performance for intrasubject analysis, although the individual differences of EEG have a greater impact on the classification of the other four models. Therefore, we conclude that LightFD can learn more features and can be better extended to the mental state detection of intersubject.

Figure 8: Classification accuracy of SVM, LMNN, and lightFD for intersubject.

In addition, similar operation with intrasubject analysis, we could get 5 groups of data with different training sets and test sets, then we also calculated and acquired the average accuracy of each of the three states using these 5 groups of data sets, which are TAV3 95.58%, DROWS 93.97%, and TAVX 83.71%. We found that the classification accuracy of TAVX is low. The reason may be that, according to statistics, the TAVX state is more likely to be misclassified into the state of DROWS.

4.3. Transfer Learning Capabilities Analysis of LightFD

Mathematically, transfer learning is defined as below [54].

Given a source domain and learning task , a target domain and learning task , transfer learning aims to help improve the learning of the target predictive function in using the knowledge in and , where , or .

Transfer learning emphasizes on the ability of a system to recognize and apply knowledge and skills learned in previous source tasks transferring to a target prediction tasks.

In this section, we try to evaluate the capabilities of transfer learning of LightFD. In particular, we wanted to measure the performance of LightFD as a general model for real time and efficient driver fatigue detection, which could be directly used for mental states identification without any additional training process. Such last feature could be very important for promoting clinical application of such EEG analysis.

As we know, there exists significant differences of EEG signals between different subjects. Therefore, it is difficult to evaluate the situation of mental states of the other subjects just from EEG characteristics of some known subjects, which means it needs to enhance the transfer learning capabilities in EEG analysis.

First, we selected EEG data from subject s1 to s9 as the training set and that of subject s10 as the test set, then we used CSP to find the projection matrix, and the rest of the operations were consistent with those mentioned in 3.1. Based on the above experimental results, we compared and analyzed the transfer learning performance of LightFD, SVM, and LMNN models, respectively.

For the identification of three mental states, namely, TAV3, TAVX, and DROWS, SVM and LMNN have the classification accuracy of 54.95% and 53.04%, respectively, whereas LightFD could reach the promising classification accuracy of 70.28%, which proves the potential of LightFD in the field of EEG analysis for transfer learning.

Furthermore, to verify the transfer learning robustness of LightFD, we conducted 10 cross-validations. Of all the ten subjects, we randomly selected two as the test set each time, and the rest as the training set. MEDA and MTLF were used for comparison with LightFD. The randomly selected testing sets for 10 cross-validations are (s5, s7), (s1, s3), (s3, s5), (s3, s7), (s6, s7), (s4, s9), (s7, s9), (s4, s8), (s5, s6), and (s6, s10), respectively, and the results are shown in Figure 9.

Figure 9: Classification accuracy of MEDA, MTLF, and LightFD for intersubject.

In the future, combining with transfer learning will be a major development trend in EEG signal processing. We believe that LightFD, a LightGBM-based model with good performance of EEG transfer learning capabilities, will bring new opportunities and progress for EEG classification and identification analysis.

4.4. Time Complexity Analysis of LightFD

In this section, to explore the feasibility of lightFD in practical applications, we analyzed and compared the time complexity of LightFD with the abovementioned 4 typical models: SVM, LMNN, CNN, and GRU.

The obvious benefit of using histogram in lightFD is that the time consumption of calculating the split gain drops from to . LightGBM usually adopts feature parallelism by vertical segmentation of samples, whereas lightFD adopts sample parallelism, namely, horizontal segmentation, to build local histogram that is then merged into full-range histogram to find the best segmentation. The communication transmission cost is further optimized from to .

The time complexity of SVM is between and , where is the number of support vectors, L is the number of training set samples, and d is the dimension of each sample (the original dimension without mapping to the high-dimensional space). In short, its time consumption depends on the matrix inversion, and the time complexity is about , where N is the number of samples. In the case of small samples, SVM can achieve the similar performance as lightFD. But as the number of samples increases, the time consumption of SVM is much higher than that of lightFD.

As a kind of distance metric learning, LMNN needs to calculate the distance between each sample and all other samples during the training process. As the number of samples increases and that of individual sample dimensions grows, it will greatly augment the time consumption of LMNN.

Deep learning models, CNN and GRU, are a kind of high-level abstraction of data by multiple processing layers composed of multiple nonlinear transformations. The complex structure determines that the time complexity is much higher than that of SVM and LMNN, although CNN and GRU tend to perform better when the sample size gets larger and the sample feature dimension becomes higher.

In CNN, the time complexity of single convolutional layer is , where M is the size of the output feature map, which is determined by four parameters such as input size X, convolution kernel size K, padding, and stride. Expressed as follows: . K is the size of the convolution kernel, is the number of input channels, and is the number of output channels. It is shown that CNN runs slower and depends heavily on the configuration of the computer under the situation of larger samples.

For GRU, the computational complexity for each update is , where K is the number of output units, C is the number of memory element blocks, S represents the size of the memory element block, H is the number of hidden units, I is the number of units that are forward connected to the memory element, the gate unit, and the hidden unit, and is the number of weights. GRU is much simpler than CNN and performs better than CNN in case of time consumption. Furthermore, it is faster than SVM and LMNN under the situation of large samples but is still inferior in time consumption than lightFD.

In summary, LightFD has a faster running speed than other traditional models on average up to 30%, which shows more outstanding performance, especially in the case of large samples, and lays the foundation for its application in real-time EEG analysis systems.

5. Conclusion

As one kind of light-weighted machine learning methods, LightFD has excellent performance in the aspects of multiclassification of EEG analysis, as well as lower time consumption, which show profound significance for practical applications. In addition, LightFD could also achieve better classification effect in the intersubject EEG classification, which suggests its potential transfer learning capability in the classification of mental states during driving by using cerebral measurements.

Data Availability

The EEG data used to support the findings of this study are restricted by the Ethics Committee of Hangzhou Dianzi University (HDU), in order to protect subject privacy. Data are available from the corresponding author for researchers who meet the criteria for access to confidential data.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This study was supported by NSFC (61671193), the National Key R&D Program of China MOST (2017YFE0118200 and 2017YFE0116800), Key Science and Technology Program of Zhejiang Province (2018C04012), International Joint Research Center for Brain-Machine Collaborative Intelligence of Hangzhou Dianzi University, and Industrial Neuroscience Lab of University of Rome “La Sapienza.” The authors also thank BrainSigns srl for the support.

References

  1. Y. Yang, Z. Yao, Z. Li, and K. Jiao, “Study on the effectiveness of EEG time-frequency induce in driving fatigue evaluation,” Mechanical Design and Manufacturing, vol. 5, pp. 94-95, 2002. View at Google Scholar
  2. S. Kar, M. Bhagat, and A. Routray, “EEG signal analysis for the assessment and quantification of driver’s fatigue,” Transportation Research Part F Psychology and Behaviour, vol. 13, no. 5, pp. 297–306, 2010. View at Publisher · View at Google Scholar · View at Scopus
  3. M. Hajinoroozi, Z. Mao, T. P. Jung, C. T. Lin, and Y. Huang, “EEG-based prediction of driver’s cognitive performance by deep convolutional neural network,” Signal Processing Image Communication, vol. 47, pp. 549–555, 2016. View at Publisher · View at Google Scholar · View at Scopus
  4. P. Aricò, G. Borghini, G. Di Flumeri, N. Sciaraffa, A. Colosimo, and F. Babiloni, “Passive BCI in operational environments: insights, recent advances, and future trends,” IEEE Transactions on Biomedical Engineering, vol. 64, no. 7, pp. 1431–1436, 2017. View at Publisher · View at Google Scholar · View at Scopus
  5. P. Aricò, G. Borghini, G. Di Flumeri, N. Sciaraffa, and F. Babiloni, “Passive BCI beyond the lab: current trends and future directions,” Physiological Measurement, vol. 39, no. 8, Article ID 08TR02, 2018. View at Publisher · View at Google Scholar · View at Scopus
  6. T. O. Zander and C. Kothe, “Towards passive brain–computer interfaces: applying brain–computer interface technology to human–machine systems in general,” Journal of Neural Engineering, vol. 8, no. 2, Article ID 025005, 2011. View at Publisher · View at Google Scholar · View at Scopus
  7. R. N. Roy, S. Bonnet, S. Charbonnier, and A. Campagne, “Mental fatigue and working memory load estimation: interaction and implications for EEG-based passive BCI,” in Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 6607–6610, IEEE, Osaka, Japan, July 2013.
  8. X.-Y. Gao, Y.-F. Zhang, W.-L. Zheng, and B.-L. Lu, “Evaluating driving fatigue detection algorithms using eye tracking glasses,” in Proceedings of the 2015 7th International IEEE/EMBS Conference on Neural Engineering (NER), pp. 767–770, IEEE, Montpellier, France, April 2015.
  9. H. Cai and Y. Lin, “An experiment to non-intrusively collect physiological parameters towards driver state detection,” Tech. Rep., SAE International, Warrendale, PA, USA, 2007, No. 2007-01-0403. View at Google Scholar
  10. J. Healey and R. W. Picard, “Detecting stress during real-world driving tasks using physiological sensors,” IEEE Transactions on Intelligent Transportation Systems, vol. 6, no. 2, pp. 156–166, 2005. View at Publisher · View at Google Scholar · View at Scopus
  11. R. N. Khushaba, S. Kodagoda, S. Lal, and G. Dissanayake, “Driver drowsiness classification using fuzzy wavelet-packet-based feature-extraction algorithm,” IEEE Transactions on Biomedical Engineering, vol. 58, no. 1, pp. 121–131, 2011. View at Publisher · View at Google Scholar · View at Scopus
  12. W. Kong, Z. Zhou, B. Jiang, F. Babiloni, and G. Borghini, “Assessment of driving fatigue based on intra/inter-region phase synchronization,” Neurocomputing, vol. 219, pp. 474–482, 2017. View at Publisher · View at Google Scholar · View at Scopus
  13. S. Hu and G. Zheng, “Driver drowsiness detection with eyelid related parameters by support vector machine,” Expert Systems with Applications, vol. 36, no. 4, pp. 7651–7658, 2009. View at Publisher · View at Google Scholar · View at Scopus
  14. R. Fu and H. Wang, “Detection of driving fatigue by using noncontact EMG and ECG signals measurement system,” International Journal of Neural Systems, vol. 24, no. 3, Article ID 1450006, 2014. View at Publisher · View at Google Scholar · View at Scopus
  15. S. Ahn, T. Nguyen, H. Jang, J. G. Kim, and S. C. Jun, “Exploring neuro-physiological correlates of drivers’ mental fatigue caused by sleep deprivation using simultaneous EEG, ECG, and fNIRS data,” Frontiers in Human Neuroscience, vol. 10, p. 219, 2016. View at Publisher · View at Google Scholar · View at Scopus
  16. C. Lin, Y. Wang, and S. Chen, “An EEG-based brain-computer interface for dual task driving detection,” Neurocomputing, vol. 129, no. 4, pp. 85–93, 2014. View at Google Scholar
  17. M. H. Alomari, A. Samaha, and K. Alkamha, “Automated classification of L/R hand movement EEG signals using advanced feature extraction and machine learning,” International Journal of Advanced Computer Science and Applications, vol. 4, no. 6, pp. 207–212, 2013. View at Publisher · View at Google Scholar
  18. G. Vecchiato, G. Borghini, P. Aricò et al., “Investigation of the effect of EEG-BCI on the simultaneous execution of flight simulation and attentional tasks,” Medical and Biological Engineering and Computing, vol. 54, no. 10, pp. 1503–1513, 2016. View at Publisher · View at Google Scholar · View at Scopus
  19. S. Charbonnier, R. N. Roy, S. Bonnet, and A. Campagne, “EEG index for control operators’ mental fatigue monitoring using interactions between brain regions,” Expert Systems with Applications, vol. 52, pp. 91–98, 2016. View at Publisher · View at Google Scholar · View at Scopus
  20. R. N. Roy, S. Charbonnier, and S. Bonnet, “Detection of mental fatigue using an active BCI inspired signal processing chain,” IFAC Proceedings Volumes, vol. 47, no. 3, pp. 2963–2968, 2014. View at Publisher · View at Google Scholar
  21. C. Zhao, C. Zheng, M. Zhao, and J. Liu, “Physiological assessment of driving mental fatigue using wavelet packet energy and random forests,” American Journal of Biomedical Sciences, vol. 2, no. 3, pp. 262–274, 2010. View at Publisher · View at Google Scholar
  22. A. Maglione, G. Borghini, P. Aricò et al., “Evaluation of the workload and drowsiness during car driving by using high resolution EEG activity and neurophysiologic indices,” in Proceedings of the 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, pp. 6238–6241, IEEE, Chicago, IL, USA, August 2014.
  23. C. Zhang, H. Wang, and R. Fu, “Automated detection of driver fatigue based on entropy and complexity measures,” IEEE Transactions on Intelligent Transportation Systems, vol. 15, no. 1, pp. 168–177, 2014. View at Publisher · View at Google Scholar · View at Scopus
  24. A. Appriou, A. Cichocki, and F. Lotte, “Towards robust neuroadaptive HCI: exploring modern machine learning methods to estimate mental workload from EEG signals,” in Proceedings of the Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems, ACM, Montreal, Canada, April 2018.
  25. Z. Yin and J. Zhang, “Cross-session classification of mental workload levels using EEG and an adaptive deep learning model,” Biomedical Signal Processing and Control, vol. 33, pp. 30–47, 2017. View at Publisher · View at Google Scholar · View at Scopus
  26. H. Aghajani, M. Garbey, and A. Omurtag, “Measuring mental workload with EEG+fNIRS,” Frontiers in Human Neuroscience, vol. 11, no. 359, 2017. View at Publisher · View at Google Scholar · View at Scopus
  27. P. Gaikwad and A. Paithane, “Novel approach for stress recognition using EEG signal by SVM classifier,” in Proceedings of the 2017 International Conference on Computing Methodologies and Communication (ICCMC), pp. 967–971, IEEE, Erode, India, July 2017.
  28. F. Wang, S. Wang, X. Wang, Y. Peng, and Y. Yang, “Design of driving fatigue detection system based on hybrid measures using wavelet-packets transform,” in Proceedings of the IEEE International Conference on Robotics and Automation, pp. 4037–4042, Hong Kong, China, May 2014.
  29. Z. Jiao, X. Gao, Y. Wang, J. Li, and H. Xu, “Deep convolutional neural networks for mental load classification based on EEG data,” Pattern Recognition, vol. 76, pp. 582–595, 2018. View at Publisher · View at Google Scholar · View at Scopus
  30. R. Hefron, B. Borghetti, C. S. Kabban, J. Christensen, and J. Estepp, “Cross-participant EEG-based assessment of cognitive workload using multi-path convolutional recurrent neural networks,” Sensors, vol. 18, no. 5, p. 1339, 2018. View at Publisher · View at Google Scholar · View at Scopus
  31. Z. Yin and J. Zhang, “Cross-subject recognition of operator functional states via EEG and switching deep belief networks with adaptive weights,” Neurocomputing, vol. 260, pp. 349–366, 2017. View at Publisher · View at Google Scholar · View at Scopus
  32. P. Thodoroff, J. Pineau, and A. Lim, “Learning robust features using deep learning for automatic seizure detection,” in Proceedings of the Machine Learning for Healthcare Conference, pp. 178–190, Los Angeles, CA, USA, August 2016.
  33. G. Ke, Q. Meng, T. Finley et al., “A highly efficient gradient boosting decision tree,” in Proceedings of the Advances in Neural Information Processing Systems, pp. 3146–3154, Curran Associates, Inc., Long Beach, CA, USA, December 2017.
  34. S. Aggarwal, L. Aggarwal, M. S. Rihal, and S. Aggarwal, “EEG based participant independent emotion classification using gradient boosting machines,” in Proceedings of the 2018 IEEE 8th International Advance Computing Conference (IACC), pp. 266–271, IEEE, Noida, India, December 2018.
  35. O. Xie, Z.-T. Liu, and X.-W. Ding, “Electroencephalogram emotion recognition based on a stacking classification model,” in Proceedings of the 2018 37th Chinese Control Conference (CCC), pp. 5544–5548, IEEE, Wuhan, China, July 2018.
  36. P. Detti, G. Z. M. de Lara, R. Bruni, M. Pranzo, F. Sarnari, and G. Vatti, “A patient-specific approach for short-term epileptic seizures prediction through the analysis of EEG synchronization,” IEEE Transactions on Biomedical Engineering, vol. 66, no. 6, pp. 1494–1504, 2019. View at Publisher · View at Google Scholar · View at Scopus
  37. C. Tan, F. Sun, and W. Zhang, “Deep transfer learning for EEG-based brain computer interface,” in Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 916–920, IEEE, Calgary, Canada, April 2018.
  38. Z. Deng, P. Xu, L. Xie, K.-S. Choi, and S. Wang, “Transductive joint-knowledge-transfer TSK FS for recognition of epileptic EEG signals,” IEEE Transactions on Neural Systems and Rehabilitation Engineering, vol. 26, no. 8, pp. 1481–1494, 2018. View at Publisher · View at Google Scholar · View at Scopus
  39. C. Yang, Z. Deng, K.-S. Choi, and S. Wang, “Takagi-Sugeno-Kang transfer learning fuzzy logic system for the adaptive recognition of epileptic electroencephalogram signals,” IEEE Transactions on Fuzzy Systems, vol. 24, no. 5, pp. 1079–1094, 2016. View at Publisher · View at Google Scholar · View at Scopus
  40. W. Zheng, Y. Zhang, J. Zhu, and B. Lu, “Transfer components between subjects for EEG-based emotion recognition,” in Proceedings of the International Conference on Affective Computing and Intelligent Interaction, pp. 917–922, Xi’an, China, September 2015.
  41. H. Zeng, C. Yang, G. Dai, F. Qin, J. Zhang, and W. Kong, “EEG classification of driver mental states by deep learning,” Cognitive Neurodynamics, vol. 12, no. 6, pp. 597–606, 2018. View at Publisher · View at Google Scholar · View at Scopus
  42. J. Wang, W. Feng, Y. Chen, H. Yu, M. Huang, and P. S. Yu, “Visual domain adaptation with manifold embedded distribution alignment,” in Proceedings of the 2018 ACM Multimedia Conference on Multimedia Conference, pp. 402–410, ACM, Seoul, South Korea, October 2018.
  43. Y. Xu, S. J. Pan, H. Xiong et al., “A unified framework for metric transfer learning,” IEEE Transactions on Knowledge and Data Engineering, vol. 99, p. 1, 2017. View at Google Scholar
  44. Y. Tran, N. Wijesuriya, M. Tarvainen, P. Karjalainen, and A. Craig, “The relationship between spectral changes in heart rate variability and fatigue,” Journal of Psychophysiology, vol. 23, no. 3, pp. 143–151, 2009. View at Publisher · View at Google Scholar · View at Scopus
  45. S. C. Segerstrom and L. S. Nes, “Heart rate variability reflects self-regulatory strength, effort, and fatigue,” Psychological Science, vol. 18, no. 3, pp. 275–281, 2007. View at Publisher · View at Google Scholar · View at Scopus
  46. T. Jung, S. Makeig, C. Humphries et al., “Removing electroencephalographic artifacts by blind source separation,” Psychophysiology, vol. 37, no. 2, pp. 163–178, 2010. View at Publisher · View at Google Scholar · View at Scopus
  47. Q. Meng, G. Ke, T. Wang et al., “A communication-efficient parallel algorithm for decision tree,” in Proceedings of the Advances in Neural Information Processing Systems, pp. 1279–1287, Barcelona, Spain, December 2016.
  48. C.-C. Chang and C.-J. Lin, “LIBSVM,” ACM Transactions on Intelligent Systems and Technology, vol. 2, no. 3, pp. 1–27, 2011. View at Publisher · View at Google Scholar · View at Scopus
  49. K. Q. Weinberger and L. K. Saul, “Distance metric learning for large margin nearest neighbor classification,” Journal of Machine Learning Research, vol. 10, pp. 207–244, 2009. View at Google Scholar
  50. Y. Lecun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, 2015. View at Publisher · View at Google Scholar · View at Scopus
  51. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778, Las Vegas, NV, USA, June 2016.
  52. O. Abdel-Hamid, A.-R. Mohamed, H. Jiang, L. Deng, G. Penn, and D. Yu, “Convolutional neural networks for speech recognition,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 22, no. 10, pp. 1533–1545, 2014. View at Publisher · View at Google Scholar · View at Scopus
  53. T. Domhan, J. T. Springenberg, and F. Hutter, “Speeding up automatic hyperparameter optimization of deep neural networks by extrapolation of learning curves,” in Proceedings of the International Conference on Artificial Intelligence, pp. 3460–3468, Phuket, Thailand, July 2015.
  54. S. J. Pan and Q. Yang, “A survey on transfer learning,” IEEE Transactions on Knowledge and Data Engineering, vol. 22, no. 10, pp. 1345–1359, 2010. View at Publisher · View at Google Scholar · View at Scopus