Research Article  Open Access
Suwicha Jirayucharoensak, Setha PanNgum, Pasin Israsena, "EEGBased Emotion Recognition Using Deep Learning Network with Principal Component Based Covariate Shift Adaptation", The Scientific World Journal, vol. 2014, Article ID 627892, 10 pages, 2014. https://doi.org/10.1155/2014/627892
EEGBased Emotion Recognition Using Deep Learning Network with Principal Component Based Covariate Shift Adaptation
Abstract
Automatic emotion recognition is one of the most challenging tasks. To detect emotion from nonstationary EEG signals, a sophisticated learning algorithm that can represent highlevel abstraction is required. This study proposes the utilization of a deep learning network (DLN) to discover unknown feature correlation between input signals that is crucial for the learning task. The DLN is implemented with a stacked autoencoder (SAE) using hierarchical feature learning approach. Input features of the network are power spectral densities of 32channel EEG signals from 32 subjects. To alleviate overfitting problem, principal component analysis (PCA) is applied to extract the most important components of initial input features. Furthermore, covariate shift adaptation of the principal components is implemented to minimize the nonstationary effect of EEG signals. Experimental results show that the DLN is capable of classifying three different levels of valence and arousal with accuracy of 49.52% and 46.03%, respectively. Principal component based covariate shift adaptation enhances the respective classification accuracy by 5.55% and 6.53%. Moreover, DLN provides better performance compared to SVM and naive Bayes classifiers.
1. Introduction
Braincomputer interface (BCI) has been one of the most interesting biomedical engineering research fields for decades. It provides a promising technology allowing humans to control external devices by modulating their brain waves. Most BCI applications have been developed for noninvasive brain signal processing which is practical to implement in realworld scenarios. There are plenty of successful EEGbased BCI applications such as word speller programs [1] and wheelchair controllers [2]. Not only can BCI be employed to mentally control devices, but also it can be implemented for understanding our mental states. Emotion recognition is one of such applications. Automatic emotion recognition algorithms potentially bridge the gap between human and machine interactions.
A model of emotion can be characterized by two main dimensions called valence and arousal. The valence is the degree of attraction or aversion that an individual feels toward a specific object or event. It ranges from negative to positive. The arousal is a physiological and psychological state of being awake or reactive to stimuli, ranging from passive to active. The valencearousal dimensional model, represented in Figure 1, of emotion is widely used in many research studies.
Electroencephalogram (EEG) is a record of the oscillation of brain electric potentials resulting from ionic current flow between brain neurons. EEG signals are acquired by measuring the electrical activities at electrode’s positions on the scalp. The 10–20 system [3] of electrode placement, illustrated in Figure 2, provides an international system to ensure standardized reproducibility. By referring to 10–20 system, a subject’s studies could be compared over time and subjects could be compared to each other. Human’s brain wave is the composition of five main frequency bands called delta (1–3 Hz), theta (4–7 Hz), alpha (8–13 Hz), beta (14–30 Hz), and gamma (31–50 Hz), as shown in Figure 3. The characteristics of each band can be utilized to estimate subject’s cognition and emotion states.
(a)
(b)
(c)
(d)
(e)
There exist several research studies, EEGbased emotion recognition systems. Koelstra et al. [5] presented methods for single trial classification using both EEG and peripheral physiological signals. Power spectrum density (PSD) of EEG signals was used as the features. A Support vector machine (SVM) classifier was used to classify two levels of valence states and two levels of arousal states. For EEG analysis results, average and maximum classification rates of 55.7% and 67.0% were obtained for arousal and 58.8% and 76.0% for valence. Soleymani et al. [6] provided a multimodal dataset, called “MAHNOBHCI,” for an analysis of human affective states. The EEG and peripheral physiological signals were employed to classify emotion states. The system used PSD of EEG signals from 32 channels as input features. A SVM classifier was implemented to classify three levels of valence states and three levels of arousal states. For EEGbased classification, the accuracy rates for valence and arousal are 57.0% and 52.4%, respectively. Huang et al. [7] developed an asymmetry spatial pattern (ASP) technique to extract features for EEGbased emotion recognition algorithm. The system employed KNearest Neighbor (KNN), naive Bayes (NB), and support vector machine (SVM) for emotion classification. The average accuracy rates for valence and arousal are 66.05% and 82.46%, respectively.
Moreover, several studies [8–11] used PSD of EEG data as the input features and performed emotion classification by using SVM. Other machine learning techniques, such as naive Bayes, KNN, LDA, and ANN, have been applied in other studies [12–15]. Although the number of research studies on EEGbased emotion recognition algorithms has been increasing in recent years, the efficiency of these algorithms is limited.
2. An Overview of Deep Learning Network
2.1. Hierarchy Feature Learning
Deep learning network (DLN) is capable of discovering unknown feature coherences of input signals that is crucial for the learning task to represent such a complicated model. The DLN provides hierarchical feature learning approach. Learned features at highlevel are derived from features at lowlevel with greedy layerwise unsupervised pretraining. This unsupervised pretraining provides the stage for a final training phase that is finetuning process with respect to a supervised training criterion based on gradient descent optimization. Consequently, the primary purpose of DLN is to learn the kind of complicated functions that can represent highlevel abstraction. A hierarchical architecture of DLN is illustrated in Figure 4.
The DLN potentially performs selftaught learning from very large numbers of sets of unlabeled data. When learning algorithms process more data, they provide better performance. The key advantage of selftaught learning and unsupervised feature learning is that the algorithm can learn from unlabeled data, and then it can learn from massive amount of information. Consequently, DLN algorithm is suitable for problems where there are a plenty of sets of unlabeled data and a handful amount of sets of labeled data.
2.2. Stacked Autoencoder
A stacked autoencoder is a neural network consisting of multiple layers of sparse autoencoders in which the outputs of each layer are wired to the inputs of the successive layers. The structure of an autoencoder is depicted in Figure 5. The autoencoder tries to learn an approximation to the identity function, shown as follows: The DLN exploits the unsupervised pretraining technique with greedy layerwise training. The algorithm performs unsupervised pretraining one layer at a time, starting from the input layer to the output layer. The first sparse autoencoder (1st hidden layer) is trained on the raw inputs to learn primary features on the inputs. During pretraining process, all of weight and bias parameters have been learned to minimize the cost function, shown in (2). Next, the algorithm performs forward propagation by using the raw inputs into this trained sparse autoencoder to obtain the primary feature activations. For pretraining in the next hidden layer, the algorithm computes its features in the same procedure from the learned features from the previous hidden layers: where is number of hidden nodes, is number of inputs, is weight of sparsity penalty, KL is KullbackLeibler divergence function, is sparsity parameter, is probability of firing activity, is weight decay parameter, and is weight of hidden nodes.
2.3. Softmax Classifier
Softmax classifier is responsible for statistically estimating the probability of output values of the DLN. Softmax classifier attempts to learn all of weight and bias parameters by using the learned features of the last hidden layer. A stacked autoencoder with 2 hidden layers and softmax classifier for binary classification is illustrated in Figure 6. In the case of binary classification (), the softmax regression hypothesis outputs , shown as follows: Softmax classifier can be generalized to be multiclass classification. The hypothesis will output a vector of estimated probabilities, shown as follows: The softmax layer needs to learn the weight and bias parameters with supervised learning approach by minimizing its cost function, shown as follows: where is number of hidden units, is number of inputs, is number of classes, is ground truth, and is weight of hidden nodes.
2.4. FineTuning Stacked Autoencoder
After completing the weight and bias parameter learning in the softmax classifier or output layer, the algorithm has to perform finetuning of all weight and bias parameters in the whole network simultaneously. Finetuning procedure treats all layers of a stacked autoencoder as a single model and improves all the weights of all layers in the network by using backpropagation technique. The standard backpropagation algorithm is used to learn the network weights and biases based on labeled training examples. The learning goal is to minimize classification errors.
2.5. DLN for EEG Data Processing
The original concept of greedy layerwise unsupervised pretraining on the deep learning networks derived from [17]. The network consisted of multilevel restricted Boltzmann machine. Later, Wulsin et al. [18] applied the unsupervised pretraining concept to a stack of autoencoder for classifying and detecting anomaly measurement in EEG waveforms. The paper demonstrated that DLNs and raw data inputs may be more effective for online automated EEG waveform recognition than other standard techniques. DLN has also been applied to classify sleep stages [19]. The study utilized an unsupervised feature learning architecture on both raw EEG data and power spectral feature extraction to perform sleep stage classification.
3. Methodology
3.1. DEAP Dataset
DEAP [20] is a multimodal dataset for analysis of human affective states. The EEG and peripheral physiological signals of 32 subjects were recorded as each subject watched 40 oneminute highlight music videos. After watching each music video, the subjects performed a selfassessment of their levels of arousal, valence, dominance, and liking. Selfassessment manikins (SAM) [21], as shown in Figure 7, were used to visualize the scales. The subjects selected the numbers 1–9 to indicate their emotion states in each category.
This study mapped the scales (1–9) into 3 levels of each valence and arousal states. The valence scale of 1–3 was mapped to “negative,” 4–6 to “neutral,” and 7–9 to “positive,” respectively. The arousal scale of 1–3 was mapped to “passive,” 4–6 to “neutral,” and 7–9 to “active,” respectively. According to the new scale mapping, the system provides 9state emotion classification: happy, pleased, relaxed, excited, neutral, calm, distressed, miserable, and depressed, shown in Figure 8.
3.2. EEG Feature Extraction
In our experiment, the proposed system employed 32channel EEG signals, without any additional peripheral physiological signals. The EEG signals were downsampled from 512 Hz to 128 Hz. The EEG channel consisted of Fp1, AF3, F3, F7, FC5, FC1, C3, T7, CP5, CP1, P3, P7, PO3, O1, Oz, Pz, Fp2, AF4, Fz, F4, F8, FC6, FC2, Cz, C4, T8, CP6, CP2, P4, P8, PO4, and O2. The power spectral density was calculated using FFT with a Hanning window of size 128 samples. The power spectral features of EEG signals on these channels were extracted in 5 frequency bands: theta (4–8 Hz), lower alpha (8–10 Hz), upper alpha (10–12 Hz), beta (12–30 Hz), and gamma (30 Hz up). In addition to the power spectral features, the difference between the spectral power of all the symmetrical 14 pairs of electrodes on the right and the left hemispheres in 5 frequency bands was extracted to measure the possible asymmetry in brain activities due to emotional stimuli. A total number of 230 EEG features were used as the input of DLN.
3.3. Feature Normalization
The baseline power was first subtracted from all of the extracted power spectral features, yielding the change of power relative to the prestimulus period, after which the features were rescaled into the range [0.1, 0.9]. This normalization process is required since the DLN uses sigmoid as the activation function in the output layer. Some of the features below and above were truncated into 0.1 and 0.9, respectively.
3.4. DLN Implementation
The proposed EEGbased emotion recognition system is implemented with a stack of three autoencoders with two softmax layers, illustrated in Figure 9. The system performs emotion classification by estimating valence and arousal states separately. Two softmax classifiers, one for valence and another for arousal, can share the outcome of unsupervised pretraining procedure because they both use the same set of unlabeled raw data. However, two softmax classifiers need to use different stacked autoencoders during finetuning backpropagation.
The DLN utilizes the unsupervised pretraining technique with greedy layerwise training, starting from the input layer to the softmax layer. The first sparse autoencoder (1st hidden layer) is trained on the inputs’ features (230 power spectral features) to learn the primary features on these input features. We use LBFGS to optimize the cost function, squared error between input features and outputs. All of parameter settings in the DLN for EEGbased emotion recognition are shown in Table 1.

Subsequently, the algorithm performs forward propagation by using the input features into this trained sparse autoencoder to obtain the primary feature activations. The features, deriving from feedforward propagation of the 1st hidden layer, must be used to perform unsupervised pretraining in the second hidden layer. The algorithm computes its features in the same procedure from the learned features from the previous hidden layers.
The weight and bias parameters of the softmax layer are trained by using a supervised learning approach. The output features of the last hidden layer are used as the input features of both softmax layers. We use a set of selfassessment emotion states (valence and arousal) of subjects as a ground truth. These softmax layers can be trained as the parameters concurrently.
After the network finishes learning weight and bias parameters in both softmax classifiers, the algorithm has to perform finetuning of all weight and bias parameters in the whole network simultaneously. However, we are not able to use the same network parameters for two classifiers. We need to save the learned parameter outcomes of unsupervised pretraining and load the parameters for finetuning process of another softmax classifier. The finetuning process treats all layers of a stacked autoencoder and softmax layer as a single model and improves all the weights of all layers in the network by using backpropagation technique with supervised approach. The backpropagation process is used to learn the network weights and biases based on labeled training examples to minimize the classification errors.
Summary of DLN training procedure is illustrated in Figure 10. The algorithm performs a greedy layerwise unsupervised pretraining process, starting from the first hidden layer to the last hidden layer. Initial weights and biases of the trained hidden layer are assigned for parameter optimizations. Next, the features from feedforward propagation of the hidden layer must be used to perform unsupervised pretraining in the next hidden layer. After finishing unsupervised pretraining in the last hidden layer, softmax training and finetuning procedures are required.
3.5. Covariate Shift Adaptation of Principal Components
Deep learning networks implemented with stacked autoencoders have capability of representing a highly expressive abstraction. Therefore, we are confronted with overfitting problems, especially with the tremendous number of input features and hidden nodes. Moreover, a nonstationary effect of EEG signal is still challenging to develop a reliable EEGbased emotion recognition. The proposed system employs the concept of principal component based covariate shift adaptation [22] to handle both overfitting problems and nonstationary effects simultaneously. Principal component analysis (PCA) [23] is to extract the most important principal components and normalize these components individually by shifting a window over the data to alleviate the effect of nonstationarity.
PCA is a statistical method that uses orthogonal transformation to convert a set of observations of possibly correlated variables into a set of values of linearly uncorrelated variables called principal components. The number of principal components is less than or equal to the number of original variables. This transformation is defined in such a way that the first principal component has the largest possible variance. The proposed system reduces the number of input features from 230 to 50 features.
To minimize the nonstationary effects of input features, the proposed system normalizes the input features with the average of previous feature values within a rectangular window of length . We performed this normalization for each input feature individually. Figure 11 illustrates the shifting window during input feature normalization for covariate shift adaptation in each video trial. In our experiments, the window size of the process is set to 10.
4. Experiments and Results
In our experiments, the efficiency of our proposed EEGbased emotion recognition system was evaluated by four experiment setups, shown in Figure 12. In the first setup, we implemented the emotion recognition by using a deep learning network with 100 hidden nodes in each layer (DLN100). We employed the feature extraction process to calculate all of input features of the DLN from 32channel EEG signals. At each epoch, the system learned 230 input features consisting of power spectral density of 5 frequency bands and the differences of power spectral densities of 14 asymmetry pairs. Next, the second experiment reduced the number of hidden nodes to 50 (DLN50) for investigating the effect of hidden node size in the DLN.
(a)
(b)
(c)
(d)
The third experiment setup, shown in Figure 12(c), exploited the PCA to alleviate overfitting problem of the DLN. The PCA extracted the 50 most important components from initial 230 input features. The extracted features were fed into the DLN with 50 hidden nodes in each layer.
The last experimental setup enhanced the efficiency of the emotion recognition system by applying covariate shift adaptation (CSA) concept to solve the problem of nonstationarity in EEG signals. The system normalized the input features with the average of previous feature values within a rectangular window of length . This normalization was processed for each input feature individually.
The classification accuracy of valence and arousal states in four experiment setups was measured with a leaveoneout cross validation scheme. The full leaveoneout cross validation of 32 subject acquisitions was performed. A training dataset was a composition of all input features from the other 31 subjects. A test dataset was the subject’s input features under evaluation. Each individual dataset consisted of power spectral features from EEG signal records while the subject was watching 40 oneminute music videos. The DLN performed its weight and bias optimization based on gradient descent approach. Therefore, the classification accuracy was occasionally affected by its initial weight and bias parameter. In our experiment, we repeated the classification accuracy measurement five times and used the average of the accuracy for further analysis.
The comparison of accuracy from four experiment setups for valence and arousal states on individual subjects is listed in Table 2. The average accuracy and standard deviation of 32 subjects in four experiments are depicted in Figure 13. The DLN100 provides the accuracy of 49.52% for valence and 46.03% for arousal. The DLN50 accuracy slightly decreases into 47.87% and 45.50%. The number of hidden nodes in the DLN affects accuracy performance of affective state classification. The greater the number of hidden nodes is, the higher accuracy the DLN provides. In experiments, the number of hidden nodes in each layer was reduced from 100 to 50 nodes. The accuracy decreased 1.62% and 0.53% for valence and arousal classifications, respectively.

There is a strong relationship between autoencoder and principal component analysis [24]. If the number of hidden nodes is less than the number of visible nodes, the autoencoder essentially performs nonlinear principal analysis (NPCA). Both approaches are responsible for learning some correlations of data. If some of the input features are correlated, then these algorithms will be able to discover some of those correlations. The PCA helps the stack of autoencoder to learn some linear correlations among the input features by acting as one more hidden layer at the input and then boost the performance of the learning task. From experimental results, the PCA increases the accuracy performance by 3.01% for valence and 3.14% for arousal.
Subsequently, we applied covariate shift adaptation (CSA) concept to alleviate the effect of nonstationarity in EEG signals. The CSA provides the classification performance to 53.42% for valence and 52.03% for arousal. The PCA+CSA setup improves the accuracy by 5.55% and 6.53% for valence and arousal states, respectively.
To evaluate the efficiency of the DLN, LIBSVM tools [25] were used to measure the accuracy performance of a SVM classifier. Its kernel function was set to radial basis function and other parameters were assigned by default values. There were three experiment setups for the SVM classifier: 230 input features, PCA, and PCA+CSA. Table 2 shows the accuracy performance of the SVM classifier.
The comparison of DLN and SVM accuracy is depicted in Figures 14 and 15 for valence and arousal states, respectively. The DLN outperforms SVM in all experiments. It is interesting to investigate the effect of PCA for feature dimension reduction. The PCA enhanced the accuracy performance of the DLN but it diminished those of the SVM. The effect of PCA on SVM is congruent with a study by Li et al. [26].
Overall accuracy of the SVM classifier to perform EEGbased emotion state classification from DEAP dataset is quite low. In our experiments, all parameters used in the SVM classifier were assigned with their default values. Moreover, the SVM exhaustedly estimated its optimal decision surfaces with a large number of sets of training data (74400 instances). These two reasons potentially lead to the SVM’s poor performance in this case.
The performance comparison among EEGbased emotion classification algorithms is shown in Table 3. We also utilized a naive Bayes (NB) classifier in WEKA tool to perform emotion state classification of the DEAP dataset with 10fold cross validation. Another NB classification technique in Chung and Yoon [16] uses a weightedlogposterior function for the Bayes classifier but its accuracy performance was measured in leaveonetrailout cross validation.
 
*Subjectdependent results [16]. 
5. Discussion
The primary purpose of this research is to explore how well the deep learning network in the version of stacked autoencoder performs EEGbased affective computing algorithm. From our experimental results, the average of emotion classification accuracy from the deep learning network with a stack of autoencoders is better than existing algorithms. Consequently, the DLN is a promising alternative as EEGbased emotion classifier. However, one of the most challenging limitations for performing EEGbased emotion recognition algorithm is coping with the problem of intersubject variations in their EEG signals.
There are several promising methods to handle the intersubject variations. Lotte and Guan [27] proposed an algorithm for learning features from other subjects by performing regularization of common spatial patterns (CSP) and linear discriminant analysis (LDA). The method regularized the estimated covariance matrix toward the average covariance matrix of other subjects. Samek et al. [28] studied transferring information about nonstationarities in data, instead of learning the taskrelevant part from others. These principal nonstationarities are similar between subjects and can be transferable. Also they have an adverse effect on classification performance, and thus removing them is favorable. We plan to implement one of these two methods, depending on the nonstationary characteristics of the dataset, for alleviating the intersubject variations in our next version of EEGbased emotion recognition system.
One of the major limitations of the DLN is its tremendous amount of computational time requirement during unsupervised pretraining and supervised finetuning procedures. In our experiment setup, the DLN for EEGbased emotion recognition is constituted of three stacks of hidden layers and each hidden layer has 100 hidden nodes. At each epoch, the algorithm learned 230 input features. To estimate an individual subject’s classification accuracy, there were in total 31 subjects watching 40 videos, each of 60 seconds ( 74,400) epochs. They are used to adjust the weight and bias parameters of the DLN. Table 1 shows other DLN’s parameter settings. The approximated time used to train the DLN is 20–25 minutes on a laptop computer (Core i53320M 2.6 GHz, RAM 8 GB, SSD 128 GB, Windows 7 64bit Professional).
To speed up training time of the DLN, we are able to exploit some parallelism between two softmax classifiers. However, we need to duplicate the stack of autoencoder implementation for valence and arousal states. Both stacks of autoencoders can be used for separated finetuning process of valence and arousal simultaneously. During unsupervised pretraining, two softmax classifiers can share the outcome of unsupervised pretraining procedure because they both use the same set of unlabeled raw data. After completing all sequences of DLN training procedure, shown in Figure 10, the DLN can be used to classify emotion states in real time. Even though the DLN requires tremendous amount of training time, it is able to perform EEGbased emotion classification in real time. During classification phase, the DLN simply feeds the input features through all layers of the network. To give better response, we are able to decrease the window size of covariate shift adaptation but we may trade off with lower classification accuracy.
6. Conclusion
The proposed EEGbased emotion recognition is implemented with a deep learning network and then enhanced with covariate shift adaptation of the principal components. The deep learning network is constituted of a stack of three autoencoders and two softmax classifiers for valence and arousal state classifications. The purpose of PCA is to reduce dimension of input features. The CSA handles the nonstationary effect of EEG signals. The classification accuracy of the DLN with PCA+CSA is 53.42% and 52.05% to classify three levels of valence states and three levels of arousal states. The DLN provides better accuracy performance compared to SVM and naive Bayes classifier. One of the major limitations for performing EEGbased emotion recognition algorithm is dealing with the problem of intersubject variations in their EEG signals. The common features of transferable nonstationary information can be investigated to alleviate the intersubject variation problems.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
References
 F. Akram, M. K. Metwally, H. Han, H. Jeon, and T. Kim, “A novel P300based BCI system for words typing,” in Proceedings of the International Winter Workshop on BrainComputer Interface (BCI '13), pp. 24–25, February 2013. View at: Publisher Site  Google Scholar
 R. S. Naveen and A. Julian, “Brain computing interface for wheel chair control,” in Proceedings of the 4th International Conference on Computing, Communications and Networking Technologies (ICCCNT '13), pp. 1–5, Tiruchengode, India, July 2013. View at: Publisher Site  Google Scholar
 F. Sharbrough, G. E. Chatrian, R. P. Lesser, H. Luders, M. Nuwer, and T. W. Picton, “American Electroencephalographic Society guidelines for standard electrode position nomenclature,” Journal of Clinical Neurophysiology, vol. 8, no. 2, pp. 200–202, 1991. View at: Google Scholar
 Wikipedia, “Electroencephalography,” March 2014, http://en.wikipedia.org/wiki/Electroencephalography. View at: Google Scholar
 S. Koelstra, A. Yazdani, M. Soleymani et al., “Single trial classification of EEG and peripheral physiological signals for recognition of emotions induced by music videos,” in Proceedings of the International Conference on Brain Informatics, Toronto, Canada, 2010. View at: Google Scholar
 M. Soleymani, J. Lichtenauer, T. Pun, and M. Pantic, “A multimodal database for affect recognition and implicit tagging,” IEEE Transactions on Affective Computing, vol. 3, no. 1, pp. 42–55, 2012. View at: Publisher Site  Google Scholar
 D. Huang, C. Guan, K. K. Ang, H. Zhang, and Y. Pan, “Asymmetric spatial pattern for EEGbased emotion detection,” in Proceeding of the International Joint Conference on Neural Networks (IJCNN '12), pp. 1–7, Brisbane, Australia, June 2012. View at: Publisher Site  Google Scholar
 G. Chanel, J. J. M. Kierkels, M. Soleymani, and T. Pun, “Shortterm emotion assessment in a recall paradigm,” International Journal of Human Computer Studies, vol. 67, no. 8, pp. 607–627, 2009. View at: Publisher Site  Google Scholar
 D. Nie, X.W. Wang, L.C. Shi, and B.L. Lu, “EEGbased emotion recognition during watching movies,” in Proceedings of the 5th International IEEE/EMBS Conference on Neural Engineering (NER '11), pp. 667–670, Cancun, Mexico, May 2011. View at: Publisher Site  Google Scholar
 X.W. Wang, D. Nie, and B.L. Lu, “EEGbased emotion recognition using frequency domain features and support vector machines,” in Neural Information Processing, B.L. Lu, L. Zhang, and J. Kwok, Eds., vol. 7062, pp. 734–743, Springer, Berlin, Germany, 2011. View at: Google Scholar
 N. Jatupaiboon, S. Panngum, and P. Israsena, “Realtime EEGbased happiness detection system,” The Scientific World Journal, vol. 2013, Article ID 618649, 12 pages, 2013. View at: Publisher Site  Google Scholar
 G. Chanel, J. Kronegg, D. Grandjean, and T. Pun, “Emotion assessment: arousal evaluation using EEG's and peripheral physiological signals,” in Multimedia Content Representation, Classification and Security, B. Gunsel, A. Jain, A. M. Tekalp, and B. Sankur, Eds., vol. 4105, pp. 530–537, Springer, Berlin, Germany, 2006. View at: Google Scholar
 O. AlZoubi, R. A. Calvo, and R. H. Stevens, “Classification of EEG for affect recognition: an adaptive approach,” in AI 2009: Advances in Artificial Intelligence, A. Nicholson and X. Li, Eds., vol. 5866 of Lecture Notes in Computer Science, pp. 52–61, Springer, Berlin, Germany, 2009. View at: Google Scholar
 G. Chanel, C. Rebetez, M. Bétrancourt, and T. Pun, “Emotion assessment from physiological signals for adaptation of game difficulty,” IEEE Transactions on Systems, Man, and Cybernetics A Systems and Humans, vol. 41, no. 6, pp. 1052–1063, 2011. View at: Publisher Site  Google Scholar
 U. Wijeratne and U. Perera, “Intelligent emotion recognition system using electroencephalography and active shape models,” in Proceedings of the 2nd IEEEEMBS Conference on Biomedical Engineering and Sciences (IECBES '12), pp. 636–641, December 2012. View at: Publisher Site  Google Scholar
 S. Y. Chung and H. J. Yoon, “Affective classification using Bayesian classifier and supervised learning,” in Proceedings of the 12th International Conference on Control, Automation and Systems (ICCAS '12), pp. 1768–1771, October 2012. View at: Google Scholar
 G. E. Hinton, S. Osindero, and Y. Teh, “A fast learning algorithm for deep belief nets,” Neural Computation, vol. 18, no. 7, pp. 1527–1554, 2006. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 D. F. Wulsin, J. R. Gupta, R. Mani, J. A. Blanco, and B. Litt, “Modeling electroencephalography waveforms with semisupervised deep belief nets: fast classification and anomaly measurement,” Journal of Neural Engineering, vol. 8, no. 3, Article ID 036015, 2011. View at: Publisher Site  Google Scholar
 M. Längkvist, L. Karlsson, and A. Loutfi, “Sleep stage classification using unsupervised feature learning,” Advances in Artificial Neural Systems, vol. 2012, Article ID 107046, 9 pages, 2012. View at: Publisher Site  Google Scholar
 S. Koelstra, C. Mühl, M. Soleymani et al., “DEAP: a database for emotion analysis; using physiological signals,” IEEE Transactions on Affective Computing, vol. 3, no. 1, pp. 18–31, 2012. View at: Publisher Site  Google Scholar
 J. D. Morris, “SAM: the selfassessment manikin. An efficient crosscultural measurement of emotion response,” Jounal of Advertising Research, vol. 35, no. 8, pp. 63–68, 1995. View at: Google Scholar
 M. Spüler, W. Rosenstiel, and M. Bogdan, “Principal component based covariate shift adaption to reduce nonstationarity in a MEGbased braincomputer interface,” EURASIP Journal on Advances in Signal Processing, vol. 2012, article 129, 2012. View at: Publisher Site  Google Scholar
 I. T. Jolliffe, Principal Component Analysis, Springer, New York, NY. USA, 1986. View at: Publisher Site  MathSciNet
 P. Baldi and K. Hornik, “Neural networks and principal component analysis: learning from examples without local minima,” Neural Networks, vol. 2, no. 1, pp. 53–56, 1989. View at: Publisher Site  Google Scholar
 C. Chang and C. Lin, “LIBSVM: a Library for support vector machines,” ACM Transactions on Intelligent Systems and Technology, vol. 2, article 27, no. 3, 2011. View at: Publisher Site  Google Scholar
 K. Li, X. Li, Y. Zhang et al., “Affective state recognition from EEG with deep belief networks,” in Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine, 2013. View at: Google Scholar
 F. Lotte and C. Guan, “Learning from other subjects helps reducing braincomputer interface calibration time,” in Proceedings of the 35th IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '10), pp. 614–617, Dallas, Tex, USA, March 2010. View at: Publisher Site  Google Scholar
 W. Samek, F. C. Meinecke, and K. Muller, “Transferring subspaces between subjects in brain—computer interfacing,” IEEE Transactions on Biomedical Engineering, vol. 60, no. 8, pp. 2289–2298, 2013. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2014 Suwicha Jirayucharoensak et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.