Computational and Mathematical Methods in Medicine

Computational and Mathematical Methods in Medicine / 2020 / Article
Special Issue

Computational Intelligence Methods for Brain-Machine Interfacing or Brain-Computer Interfacing

View this Special Issue

Research Article | Open Access

Volume 2020 |Article ID 8303465 | https://doi.org/10.1155/2020/8303465

Ayan Seal, Puthi Prem Nivesh Reddy, Pingali Chaithanya, Arramada Meghana, Kamireddy Jahnavi, Ondrej Krejcar, Radovan Hudak, "An EEG Database and Its Initial Benchmark Emotion Classification Performance", Computational and Mathematical Methods in Medicine, vol. 2020, Article ID 8303465, 14 pages, 2020. https://doi.org/10.1155/2020/8303465

An EEG Database and Its Initial Benchmark Emotion Classification Performance

Guest Editor: Yi-Zhang Jiang
Received06 Jun 2020
Accepted06 Jul 2020
Published03 Aug 2020

Abstract

Human emotion recognition has been a major field of research in the last decades owing to its noteworthy academic and industrial applications. However, most of the state-of-the-art methods identified emotions after analyzing facial images. Emotion recognition using electroencephalogram (EEG) signals has got less attention. However, the advantage of using EEG signals is that it can capture real emotion. However, very few EEG signals databases are publicly available for affective computing. In this work, we present a database consisting of EEG signals of 44 volunteers. Twenty-three out of forty-four are females. A 32 channels CLARITY EEG traveler sensor is used to record four emotional states namely, happy, fear, sad, and neutral of subjects by showing 12 videos. So, 3 video files are devoted to each emotion. Participants are mapped with the emotion that they had felt after watching each video. The recorded EEG signals are considered further to classify four types of emotions based on discrete wavelet transform and extreme learning machine (ELM) for reporting the initial benchmark classification performance. The ELM algorithm is used for channel selection followed by subband selection. The proposed method performs the best when features are captured from the gamma subband of the FP1-F7 channel with 94.72% accuracy. The presented database would be available to the researchers for affective recognition applications.

1. Introduction

Emotion is a mental condition [1] of a person. It is directly or indirectly connected to high intensity and high hedonic content namely, thoughts, feelings, behavioral responses, and a level of happiness or unhappiness [2]. The objective of emotion recognition is to identify a human’s temporal emotional state based on facial expressions and verbal expressions automatically. Multimedia elements such as texts, audio, and video are also used frequently to induce human expressions for emotion recognition [3]. Body movements with gestures and speech are also combined with facial expressions to classify emotion more accurately [4]. Unavoidably, feelings play a significant role in making relations with others. It also helps to increase interaction with computers. Affective computing is a branch of study, where human interaction with computers using emotion and its applications are taught [5]. Affective computing finds its applications spanning in various areas like medicine, e-learning, monitoring, marketing, entertainment, and law. One can find some real-life examples of affective computing such as counseling and determining client’s medical state, determining patients sentiment and relief level about the treatment during healthcare, analyzing the emotion and thereby adjusting the learning technique and presentation according to the style of a learner in the case of e-learning, remote health checkup of senior citizens, expanding means of communication and assisting the people with autism, and determining fatigue in the case of driving and alerting in advance. Since the emotional state of a human may affect attentiveness, problem-solving, and policy-making skills, the vision of affective computing is to build systems that are capable to identify and influence a person’s emotion to increase yield and effectiveness of working with computers. Electroencephalography (EEG) is an electrophysiological checking strategy to record the electrical activity of the human brain. It is ordinarily noninvasive, with the electrodes placed along the scalp. EEG estimates voltage variations because of ionic current within the neurons of the cerebrum. The state-of-the-art emotion recognition methods [6, 7] can be broadly divided into three types namely, knowledge-based methods, statistical approaches, and hybrid techniques [8]. In knowledge-based, emotion detection is carried out by utilizing domain understanding and the semantic and syntactic features of language [9] using WordNet, SenticNet [10], ConceptNet, and EmotiNet [11]. However, it is unable to handle concept nuances and complex linguistic rules [8]. In statistical techniques, generally different supervised machine learning algorithms such as Support Vector Machines [12], Naive Bayes, and Maximum Entropy [13] are considered in emotion classification using a large set of annotated data. The classification accuracy is higher in the case of statistical methods, and it depends on the sufficiently large training set. Few works consider both knowledge-based techniques and statistical methods in emotion identification by exploiting complementary features [14, 15]. The classification accuracy provided by hybrid approaches is higher as compared to the previous two methods. However, it is computationally more challenging [9]. Nowadays, researchers are using physiological signals such as EEG and electrocardiogram (ECG) signals for emotion recognition [16, 17]. The different sources of physiological signals in the human body are shown in Figure 1.

However, when compared to emotion portrayed through facial expressions, emotion captured through EEG signals is more reliable because someone can fake his/her facial expressions; in other words, he/she may feel a different emotion from within and act and express something else, or in some cases, he/she might not be able to express through gestures or facial expressions the feeling they are experiencing inside. Interested readers can go through [19] for knowing other advantages of EEG signals over facial expression in emotion recognition. Data is an integral component of any emotion recognition task. It is rare to get publicly available annotated data which is essential for training and testing of machine learning algorithms. Some of the publicly available multimodal sources of data in the form of multimedia contents such as texts, audio, videos, or physiological signals to perform emotion recognition research are briefly described as follows: Belfast facial expression database [20] was created mainly for the investigation of sex differences, cultural differences, and individual differences in the expression and perception of a wide range of emotions using short TV programs and interview recordings [20]. HUMAINE facial expressions database [21] consists of almost all audio-visual recordings in diverse emotion-rich scenarios. SEMAINE is also an audio-visual recording [22] between a person and a virtual assistant and comprises of annotated emotion like angry, happy, fear, disgust, sadness, contempt, and amusement. IEMOCAP is also audio-visual recordings of dyadic sessions between actors and consists of emotion annotations such as happiness, anger, sadness, frustration, and neutral state [23]. eNTERFACE considers pleasure, irritation, sorrow, amazement, disgust, and panic in the form of audio-visual recordings from seven nationalities [24]. DEAP [16] and DREAMER [17] databases consist of EEG, ECG, and face video recordings [16]. This database is built, while people were watching film clips. All the data are annotated properly in terms of valence, arousal, and dominance of people. In [25], Alakus et al. offered an EEG database called GAMEEMO for various games.

It is clear from the literature that a few publicly available EEG signals databases exist. So, the salient contributions of the proposed study are as follows: (i)A new database is presented, which consists of four emotions namely, happy, sad, neutral, and fear in the form of EEG signals of 44 subjects(ii)The captured EEG signals are classified based on discrete wavelet transform (DWT) and extreme learning machine (ELM).(iii)The ELM algorithm is used for channel selection followed by subband selection

The rest of the work is as follows: the background study related to the human brain and emotion is discussed in Section 2. Section 3 presents a new database for affective computing. Materials and methods used for noise removal, feature extraction, and classifier are discussed in Section 4. Section 5 presents the experimental results. Finally, the conclusion is drawn in Section 6.

2. Background Study

In the following section, we give a brief introduction of the human brain and the lobes present in it which are responsible for emotion and also about the characteristics of EEG signals.

2.1. Human Brain

The brain is the command center for the human nervous system. It is responsible for all of the body’s activities. It receives sensory input from the many sensory receptors. Then, it processes, interprets, and decides what actions should be taken. It also sends the output to the muscles. The brain has three main parts viz., cerebellum, brain stem, and cerebrum. The cerebellum accepts data from the many sensory receptors, the spinal cord, and other parts of the brain and then manages motor movements. It also involved involuntary movements such as balance, posture, coordination, and speech, governing smooth muscular function. The brain stem manages the movement of messages between the brain and the other parts of the body. The basic body activities like breathing, swallowing, heart rate, blood pressure, and consciousness are managed by stem. On the other hand, the cerebrum is the largest part of the human brain, and it is further divided into hemispheres viz., left and right. Each hemisphere has four lobes namely, frontal, temporal, occipital, and parietal as in Figure 2. The cerebrum is mainly responsible for the action and thought. The various regions of cerebrum control various actions as described below: (i)The frontal region is responsible for the movement of the body, speech, concentration, processing memories from the limbic region, emotional reaction, problem-solving, the meaning of words, personality, and planning. In the frontal lobe region, neurons of dopamine-sensitive type are located which help with the feelings of attention, memory, reward, pleasure, and planning. It contains a part of the limbic region which is responsible for some basic emotions as stated above(ii)The temporal region is responsible for hearing, emotion, and regulate emotions like hunger, thirst, recognizing faces, and long-term memory which are responsible for storing, forming, and processing visual memory. It contains some part of the limbic region which is responsible for some basic emotion as stated above(iii)The occipital region is responsible to evaluate the distance, depth, size, helps in color recognition, interprets the visual stimuli, and transmits the visual data to the other regions of the brain and other regions respond to this information(iv)The parietal region is responsible to feel the sensations of touch, process, taste, and temperature

2.2. Emotions

Emotion is a state of feeling and change in both physical and psychological influence of our behavior which we people experience in our daily routine, and it comprises three components namely, abstract understanding, a physiological reaction, and a social or expressive reaction [26, 27]. Emotion has been portrayed as discrete and steady reactions to occasions outside or inner with hugeness for the life form [28]. Emotion plays an important role in human-human communication. Emotion is often interconnected with several factors like mood, personality, temperature, one’s circumstances, or relationships with others. The emotional state can be expressed through two dimensions valence and arousal. Valence is the dimension that measures positive or negative effects due to emotion. Arousal is the dimension that measures how soothing or exciting the emotion is, and it is represented in terms of high and low. There are two different aspects of representing emotion. The first aspect shows that essential feelings have advanced through regular choices. Plutchik suggested eight primitive emotions anger, fear, sadness, disgust, surprise, curiosity, acceptance, and joy [29]. In the second aspect, given perception, the feelings are mapped into the valence, arousal, and dominance measurements. Valence goes from extremely positive sentiments to exceptionally negative or unpleasant to joy; excitement goes from states like drowsy to energized; lastly, predominance compares to the quality of the feeling [30]. A considerable lot of our regular exercises rely upon our emotional state. Assessing emotion is the key to understanding people. Consequently, human emotion acknowledgment is vital for the precise classification of human emotion. We will in this manner center around the strategies utilized for recognizing emotion.

As human-human association happens, here come brain-computer interface (BCI) systems. For a technologically invaded life, the brain-computer interaction system plays an important role. As machines impacted our lives from numerous points of view in the previous decade, machines ought to be able to find, translate, get to, and coordinate the emotional states of humans during an interaction. This human-machine association is made simple with BCI frameworks. Nowadays, these BCI frameworks have the capacity for emotion recognition. To have an effective emotion acknowledgment, a huge assortment of strategies like emotion recognition from outward appearances, the pitch of voice, signals from self-ruling sensory system, signals from EEG.

3. EEG Signals Database for Affective Computing

A very limited number of EEG signals databases are observed in the literature. This motivates us to develop an EEG signals database for research purposes. This section presents the different protocols that are considered while capturing signals of a human. The detailed description of each protocol is discussed in subsequent sections.

3.1. EEG Sensor

EEG is a therapeutic imaging procedure that interprets scalp electrical movement produced by cerebrum structures, i.e., it measures voltage variances coming about because of ionic current streams inside the neurons of the cerebrum. EEG machine used here is EEG traveler braintech 32+ CMEEG-01, which is of 32 channels. This model has many characteristics such as the most friendly used tool-bar, a facility of deleting events and comments, online/offline reformatting of gain, filter, and sweep speed. It also had EEG flash with high-intensity LED with a built-in rechargeable Li-ion battery. It is a highly advanced database management with different searching options. It operates with computer USB power.

Some of the specifications of this EEG traveler are it has 32 input channels and 32 display channels. The frequency band range is from 0.1 Hertz (Hz) to 100 Hz. The maximum impedance limit is . The sampling frequency is 1024 Hz for at ADC and 256 Hz for internal storage. It has a high-cut digital filter of the range 0.5 Hz to 100 Hz and low-cut digital filter of the range 0.1 Hz to 7 Hz. There is EEG paste which is used to adhere to the 32 electrodes or channels to the scalp for recording signals. A piece of EEG setup utilized for recording appears in Figure 3.

3.2. EEG Signals

EEG signals are principally used to track the action of the human brain utilizing electrical signals. The identification of emotional changes is normally made simple through EEG signals. Along these lines, we have picked this EEG flag-based approach for classification of human emotion to comprehend the essential feelings like happy, fear, sad, and neutral. The EEG recording is carried out with the assistance of the electrodes that are associated with the CLARITY EEG setup. These electrodes are little and level metallic discs that are joined to the scalp utilizing EEG gel/paste. For each action of the human body, cerebrum cells communicate with each other and these exercises of the brain will be recorded in a computer with the assistance of electrodes, and it is estimated in small scale voltage contrast. Voltage contrasts estimated at electrodes are little so they are digitized send to enhancer, and we can see the amplified data in terms of voltage. The EEG signals are helpful for doctors to diagnose disorders like a brain tumor, head injury, and encephalitis. This EEG signal is additionally used to measure the bloodstream level in the cerebrum all through the careful procedure and to screen the brain activity of the individual in unconsciousness/coma. EEG is considered a protected technique. It will not cause any distress or any sensation. It is not ok for a person who is experiencing seizure issue or epilepsy, since it might cause seizure during the procedure as blazing of light on the face, deep breathing is associated with the procedure. It might cause different issues relying upon the medical state of the person experiencing EEG recording.

3.3. Brainwaves

Generally, five types of electrical patterns or brain waves have been observed simultaneously by the researchers across the cortex throughout a day in our awaking state or even when we are asleep. These waves in order of lowest frequency to highest are as follows: delta (0-4 Hz), theta (4-7 Hz), alpha (8-13 Hz), beta (13-30 Hz), and gamma (30-63 Hz). It is observed that the range of these waves fluctuates a few Hz among various studies. These waves jointly perform mental functioning effectively and efficiently. However, each brain wave has a unique role. One particular brain wave will be dominant at a time while performing a particular task, and it depends on the state of cognizance that a person is in. If one of the five types of brain waves is either over synthesized and/or under synthesized in our brain, it can cause problems. The detail description of each of the useful brain wave is as follows: (i)Delta waves are related to the deepest levels of pleasure and therapeutic rest, healing sleep. These are the slowest captured brain waves in people. Generally, higher levels are noticed in infants and kids. These are produced in a lesser amount as one grows old even during deep sleep. These are also observed too many of our unconscious bodily functions such as regulating heartbeat and digestion. Adequate levels of delta waves help to feel jubilant after we wake up refreshed. However, irregular delta activity may hamper normal learning ability or maintaining awareness(ii)Theta waves are linked in daydreaming and sleep. These are linked to us experiencing and feeling deep and raw emotions. However, higher levels of theta activity may result in an episode of depression and may make people “highly suggestible” according to the fact that they are in a deeply relaxed, semihypnotic state. These waves help to improve our intuition, creativity, and make us feel more natural. It is also involved in therapeutic rest. The range of these waves is very helpful until not generated in excess during waking(iii)Alpha waves establish a “frequency bridge” between our cognizant reasoning (beta) and intuitive (theta) mind. These are known to enable quiet us to down and advance sentiments of deeper relaxation and substance. Beta waves assume a functioning part in arranging coordination and correspondence and do not happen until three years of age in person. Higher levels of beta activity and lower levels of alpha activity may result in an emotional state called “alpha blocking.”(iv)Beta waves are the low-amplitude high-frequency brain waves that are generally found in an awake person. These are produced when a person is incognizant in states like cognitive reasoning, calculation, reading, speaking, or thinking. Higher levels of beta waves may result in stress and anxiety. The beta activity of a person will increase when he or she drinks a stimulant called caffeine that is found especially in tea or coffee(v)In the area of neuroscience, a more recent discovery of brainwaves is gamma waves. Researchers are constantly exploring to know more about gamma waves. They have examined a strong relation between meditation and gamma waves. In addition to a healthy cognitive task, these perform more complex tasks. Gamma waves are important for learning, memory, and information processing. It is thought that the 40 Hz gamma wave is important for the binding of our senses in regard to perception and are involved in learning new information. Lower levels of gamma activity are observed in people with mentally challenged

3.4. EEG Electrode Positioning

The recordings of the EEG signals are performed according to the international standard 10/20 electrode system [31] as shown in Figure 4. Figure 4 depicts the 10/20 international standard electrode system in which the distance between any two adjoining electrodes can be either 10% or 20% of the total front to back or right to the left of the skull. This 10/20 system is based on the connection between the position of the electrode and base zone of the cerebral cortex. Every electrode is demonstrated with a letter and a number which is utilized to recognize the lobe and hemisphere where it must be placed (F, T, P, C, and O letters alludes to frontal, temporal, parietal, central, and occipital lobes separately). There are numbers from 1 to 8 given to the electrodes among which 1, 3, 5, and 7 stand for the electrodes to be placed on the left hemisphere and 2, 4, 6, and 8 stand for the electrodes to be placed on the right hemisphere. “Z” refers to the electrode that is to be placed on the midline. Besides these electrodes, we also have ground and reference electrodes. Four anatomical reference points on head are utilized for the outright position of electrodes specifically nasion, inion, and both preauricular focuses. Nasion is the middle of forehead and nose, inion is the point from the posterior of the head which is the lowest of the skull, and preauricular focuses are those which are front to ear [32].

3.5. Impedance Check

After placing all the electrodes in their desired positions, we should check the impedance, which is whether all the electrodes are connected properly or not. If not, we should again check those electrodes which are not connected properly and connect them. The impedance is verified in the software which is provided with the CLARITY EEG machine. During the recording procedure, the impedance of the electrodes should not exceed the maximum impedance limit. The impedance corresponding to each electrode is shown in Figure 5. If the limit exceeds the maximum impedance level, color shown in the impedance check option of the software used to record the EEG signals changes from green to red. During this impedance check, the recording is stopped.

3.6. Stimuli

The stimuli selected in this experiment through various steps. First, we should consider the different types of emotion, and among them, we should select some which we want for classification. Here, four emotions namely, sad, happy, fear, and neutral are considered. Total stimuli have 12 videos, which means each emotion has 3 videos. Each video is of approximately 150 seconds which is 2 minutes and 30 seconds [33]. After watching every video, the person should relax for 15-20 seconds for neutralizing the emotion that he had felt during that period. During relaxing for a certain period, the person should close his/her eyes and relax to avoid the surroundings and prevent disturbances around them. The signals have been recorded in a close room with no lights and room temperature is 25-30 degrees Celsius. We have selected the stimuli of our required emotion from featured films based on the criteria like comedy scenes for happy emotion and emotional scenes for sad motion and scary scenes for fear emotion and neutral scenes for neutral emotion. There should be no noise or disturbance around the one who is watching the stimuli because, if any disturbance has taken place, the person might get deviated from the emotion, and the signals during that period of disturbance will be useless. The snapshot of happy stimuli, the reaction of a subject while watching that stimuli, and the captured EEG signals of four channels namely, FP2-F4, FP2-F8, FP1-F3, and FP1-F7 are shown in Figure 6. Figure 7 shows the snapshot of sad stimuli, the subject’s reaction, and the acquired EEG signals, respectively. The snapshot of fear stimuli, the reaction of a subject while watching that stimuli, and the captured EEG signals are shown in Figure 8. Figure 9 illustrates the snapshot of neutral stimuli, the subject’s reaction, and the acquired EEG signals, respectively. The horizontal and vertical axes of Figures 6(c), 7(c), 8(c), and 9(c) are mean sample sizes and amplitudes of EEG signal, respectively.

3.7. Pattern Matching

These 12 videos are divided into 3 patterns in which each pattern consists of 4 videos of a different emotion. As the person after watching the first round of the pattern, he needs to match the videos according to the emotion provided on the left side. We should take care of creating these patterns while arranging the videos, because no pattern should be the same as of the previous one.

Pattern 1: neutral–sad–fear–happy (Figure 10).

Pattern 2: sad–happy–neutral–fear (Figure 11).

Pattern 3: happy–neutral–sad–fear (Figure 12).

This is because of the person who is matching the pattern should remember the four videos and match the pattern as to how he felt. For better results, the patterns should be varied. It is because if we change pattern we can make sure that the person who is matching the pattern is matching as how he felt rather than remembering the previous pattern and matching as it is in the new one.

3.8. Database Summary

No. of videos: 12. Video duration: 150 seconds (approximately). Selection method: manually. Total no. of participants: 44 ().

4. Materials and Methods

A schematic block diagram of the proposed system for classifying emotion is displayed in Figure 13. The proposed method includes EEG signal acquisition, removal of noise and division of subbands, extraction of features, and classification. EEG signals acquisition procedure is presented in Section 3. The rest of the steps would be described in the subsequent sections.

4.1. Discrete Wavelet Transform

It is clear from Section 3.3 that EEG signals consist of mainly five types of brain waves. However, EEG signals also include noise. It is an unwanted signal that corrupts the signal of interest. So, the noise has to be removed from the raw signals to extract useful information. Thus, the first step in the processing of EEG signals is the filtering of raw signals. EEG signal is nonstationary, and multiresolution DWT is a widely used decomposition method for eliminating noise from a nonstationary signal [26]. The frequency of the EEG signal that we have captured is 256 Hz. Daubechies 8 DWT (db8) is adopted here to decompose raw EEG signal into detail coefficients and approximation coefficients up to level. So, a total of 6 detail subbands (D1 to D6) and 1 approximation subband (A6) would be generated. These subbands are also called waves. Here, D1, D2, D3, D4, D5, D6, and A6 subbands are known as noise (125-256 Hz), noise (63-125 Hz), gamma (30-63 Hz), beta (13-30 Hz), alpha (8-13 Hz), theta (4-7 Hz), and delta (0-4 Hz), respectively. The various subbands obtained using “db8” DWT of a randomly selected channel from neutral emotion are shown in Figure 14.

4.2. Extraction of Features

The extraction of features is a process of unveiling hidden characteristics from the input signal. In other words, a set of features can represent an input signal. Moreover, this feature set illustrates a specific behavior or pattern depicted by the original input signal. Every signal is normalized to zero mean and unit variance. Ten popularly used features are extracted from each subband for emotion classification considered in this work, which is as follows: (i)Average amplitude changeAverage amplitude change is the mean value of the changes between two consecutive sample values. The value of AAC is estimated using Eq. (1). where is the sampling point of an EEG signal and C is the total number of sample points. In our case, the value of C is 2560 because the sampling frequency of the device used is 256 Hz, and the duration of the signal is 10 s. It is also referred to as difference absolute mean value.(ii)ActivityActivity is the measure of the mean power which is calculated by Eq. (2). It also represents the activeness of the signal. It also represents the variance of the EEG signal. where represents the variance of an EEG signal.(iii)Absolute square root sumThe absolute square root sum is calculated using Eq. (3) which is the sum of the square root of each sample point in a signal. where is the total number of sample points.(iv)Clearance factorThe clearance factor is calculated through Eq. (4), and it is a ratio of the peak value of a signal to the square of the average of the square root of the absolute value of the EEG signal. where peak value is obtained using .(v)Root mean squareRoot mean square (RMS) is defined as the square root of the average of the square of each sample point of a signal which is estimated using Eq. (5). (vi)Crest factorCrest factor is the ratio of the peak value of the signal to the RMS value, and it is found using Eq. (6). (vii)Shape factorThe shape factor is the ratio of RMS value to the average of the absolute value of the signal which is computed by Eq. (7). (viii)Log detectorThe nonlinear detector is defined as Eq. (8). (ix)MobilityMobility is approximated from Eq. (9) which is defined as the square root of the ratio of the variance of the first derivative of signal to the variance of the signal. (x)Absolute sumAbsolute sum is defined by Eq. (10). Every machine learning algorithm is associated with some objective functions. Sometimes, these objective functions may not work properly without normalization, because the range of values of raw data varies greater than usual. For example, most of the classifiers estimate the distance between two points by the Euclidean norm. If one of the features or attributes has a wide range of values, the distance will be governed by this specific feature. Thus, the range of all features should be normalized so that each feature contributes approximately proportionately to the final distance [34]. So, normalization is done on extracted features also.

4.3. Extreme Learning Machine Classifier

The ELM is an efficient single-layer feedforward neural networks [35]. It is mainly considered as a classifier, regressor, and cluster analysis. However, it is adopted here for emotion classification because of the better generalization, controllability and robustness, and fast learning rate. It can solve the overfitting problem as compared to other conventional neural networks. It uses the empirical risk minimization theory. Moreover, it averts local minimization and multiple iterations. Furthermore, the learning process of the ELM takes a single iteration only. Finally, tune is not required for updating the parameters of hidden nodes. These nodes can be acquired from their forefathers without any modification or can be randomly assigned. Normally, the outcome weights of hidden nodes are gained in a single step. Let us consider an ELM with a single hidden layer and the output function of the hidden node is , where f is a sigmoid function and and are the parameters of the hidden node. The output function of the ELM for single-layer feedforward network with hidden nodes is expressed by Eq. (11). where is the output weight of the hidden node. The outcome of the hidden layer is estimated using Eq. (12).

The hidden layer yield matrix of ELM is expressed by Eq. (13), when the number of training samples is .

The training data target vector is represented by Eq. (13).

The objective function of ELM is shown in Eq. (14). where and , The basic form of ELM algorithm considered in this work is as follows: where and are the two matrices used to denote weights of input to hidden layer and hidden to output layer, respectively, and is the sigmoid activation function. The steps of the ELM algorithm are as follows: where , , and denote feature matrix, response variables, and pseudoinverse, respectively. Scholars use the ELM due to the property of low computational complexity. Most training can be accomplished in minutes, and even in some cases, the ELM takes seconds to complete the training process. The complexity of the ELM can be represented by , where and denote input weights and biases, respectively [36].

Step 1. is filled by some random values.

Step 2. is estimated as per Eq. (17).

5. Experimental Results and Discussions

All the programs are implemented in Matlab 2017a, and these programs are executed on a laptop of the following specification: Intel(R) Core(TM) i5 5200U CPU @ 2.20 GHz, 4 GB RAM, 64-bit Windows Operating System.

At the very first step, DWT takes noisy EEG signals as inputs and produces noise-free EEG signals in the form of five subbands as outputs. Every subband is normalized to zero mean and unit variance. Ten features are extracted from these subbands in the second step. Then, all the features are normalized in the range of 0 and 1. In this study, 44 volunteers and their four emotions are considered for classification. Three videos are captured for each emotion. Ten pages of information for each video are stored while capturing videos. Four channels namely, FP2-F4, FP2-F8, FP1-F3, and FP1-F7 are selected out of 32 channels as these channels are placed in the frontal region of the head, and the frontal region is mainly responsible for identifying emotion correctly. Each channel has 5 subbands after noise removal. So, we have samples altogether for each emotion. It means we get a feature matrix of size , where 26400 is the sample size and 10 features are there for each sample. Three experiments are performed in this study. In the first experiment, channel-wise classification is carried out to know which channel is more responsible for emotion recognition. Several samples of each channel would be equal to , and each channel will have 10 features. The feature matrix is divided into two sets namely, training and testing sets. The former one has 70% of total samples, and the rest of the samples are part of the testing set. Training samples and their class labels are fed into ELM for building a trained model. Finally, the test sample would be provided one after another as input and predicts its class label as output with the help of the trained model. Five measures namely, accuracy, specificity, sensitivity/recall, precision, and f1-score [3739] are considered in this study for measuring the emotion recognition performance of each channel. The obtained value of each measure is reported in Table 1. It is clear from Table 1 that the performance of the first channel, i.e., FP1-F7 is better than the other three channels as the values of accuracy, sensitivity, specificity, precision, and f1-score are 94.72, 98.24, 94.72, 94.71, and 94.71, respectively, which are the highest among the four channels.


FP2-F4FP2-F8FP1-F3FP1-F7

Accuracy90.8591.2394.0194.72
Specificity96.9597.0798.0398.24
Sensitivity90.8591.2394.0194.72
Precision90.8291.0294.0994.71
F1-score90.8291.2094.0994.71

Subband wise emotion classification is performed in the second experiment. The number of samples of each subband would be equal to . So, the size of feature matrix would be . Then, 70% of the total samples are allocated for training. The remaining 30% samples are referred to as the testing set. ELM is trained by the training samples and their class labels. Test set is used then to predict the class labels of the testing samples. Accuracy is computed based on actual and predicted classes, and subband wise performance accuracy is reported in Table 2.


ChannelsAlphaBetaGammaThetaDelta

FP2-F486.5388.0190.8573.8684.05
FP2-F885.7988.2191.2380.6785.95
FP1-F392.3491.0994.0175.9383.73
FP1-F794.7094.2394.7276.4486.16

It is observed from Table 2 that the gamma subband provides the highest overall accuracy among all the subbands. Here, no emotion is considered separately. So, in the final experiment, only features are extracted from the gamma subband of the FP2-F4 channel. The number of samples would be equal to for each emotion. Again, the same procedure is adopted for classifying emotion by ELM. Emotion-wise classification performances are shown in Table 3. Table 4 reports the classification performance that is achieved with the help of features extracted from the gamma subband of the FP2-F8 channel. Similarly, the recognition performances are reported in Tables 5 and 6 only when features are obtained on gamma subband of the FP1-F7 channel.


HappySadNeutralFear

Accuracy95.7193.6394.1597.68
Specificity96.8396.0196.2998.13
Sensitivity92.5986.7688.0096.48
Precision91.2688.3089.1794.98
F1-score91.9287.5288.5895.72


HappySadNeutralFear

Accuracy96.0794.8593.9296.95
Specificity97.4296.8595.5498.38
Sensitivity92.3190.0289.0893.14
Precision92.7590.1887.0195.57
F1-score92.5390.1088.0394.34


HappySadNeutralFear

Accuracy98.3395.2594.7299.74
Specificity98.7296.7796.5999.75
Sensitivity97.2290.7489.1799.72
Precision96.4290.5089.7899.30
F1-score96.8290.6289.4799.51


HappySadNeutralFear

Accuracy97.9595.8695.7599.69
Specificity96.8597.5697.1699.87
Sensitivity97.4290.8591.5999.18
Precision94.9592.6691.5999.61
F1-score96.1791.7591.5999.39

6. Conclusion

Classification of human emotion using EEG signals is the best possible method, because it provides the actual neurophysiology of brain states. In this work, DWT decomposes each EEG signal into subbands which are further used for extraction of features from that subbands. The channel-wise extracted features are fed as inputs to the ELM classifier for the classification of happy, fear, sad, and neutral emotion. This method provides better results using a single EEG channel. The single-channel-based emotion classification ability of the proposed method shows its adroitness for the practical applications of emotion classification in BCI systems. Recent state-of-the-art methods would also be considered to compare our proposed method. Deep learning convolution neural network may be employed to denoise the EEG signals followed by classification emotions. Here, we mainly concentrated to capture EEG signals database. We are also planning to consider more number of subjects to extend the database. We will focus to make a multimodality based BCI system for emotion recognition using face biometrics and physiological signals such as EEG signals, blood pressure, respiratory rate, body temperature, and galvanic skin response in near future.

Data Availability

The “EEG signals related to emotions” data used to support the findings of this study are currently under embargo, while the research findings are commercialized. Requests for data, 6 months after publication of this article, will be considered by the corresponding author under license agreement.

Conflicts of Interest

The authors declare no conflict of interest.

Acknowledgments

This work is partially supported by the project “Prediction of diseases through computer assisted diagnosis system using images captured by minimally-invasive and non-invasive modalities”, Computer Science and Engineering, PDPM Indian Institute of Information Technology, Design and Manufacturing, Jabalpur India (under ID: SPARC-MHRD-231). This work is also partially supported by the SPEV project “Smart Solutions in Ubiquitous Computing Environments”, 2020, University of Hradec Kralove, Faculty of Informatics and Management, Czech Republic (under ID: UHK-FIM-SPEV-2020-2102) and by the Research University Grant Vot-20H04 at Universiti Teknologi Malaysia(UTM), Malaysia Research University Net-work (MRUN) Vot 4L876. We would also like to thank you for consulting to Sebastien Mambou and Ayca Kirimtat, Ph.D. students at FIM UHK.

References

  1. J. Panksepp, Affective Neuroscience: The Foundations of Human and Animal Emotions, Oxford University Press, 1998.
  2. D. L. Schacter, D. T. Gilbert, and D. M. Wegner, Introducing Psychology, Macmillan, 2009.
  3. S. Poria, E. Cambria, R. Bajpai, and A. Hussain, “A review of affective computing: from unimodal analysis to multimodal fusion,” Information Fusion, vol. 37, pp. 98–125, 2017. View at: Publisher Site | Google Scholar
  4. G. Caridakis, G. Castellano, L. Kessous et al., “Multimodal emotion recognition from expressive faces, body gestures and speech,” in IFIP International Conference on Artificial Intelligence Applications and Innovations, pp. 375–388, Springer, 2007. View at: Google Scholar
  5. A. Kolakowska, A. Landowska, M. Szwoch, W. Szwoch, and M. R. Wrobel, “Emotion recognition and its applications,” in Human-Computer Systems Interaction: Backgrounds and Applications 3, pp. 51–62, Springer, 2014. View at: Google Scholar
  6. G. Chen, X. Zhang, Y. Sun, and J. Zhang, “Emotion feature analysis and recognition based on reconstructed eeg sources,” IEEE Access, vol. 8, pp. 11907–11916, 2020. View at: Publisher Site | Google Scholar
  7. Y. Luo, Q. Fu, J. Xie et al., “Eeg-based emotion classification using spiking neural networks,” IEEE Access, vol. 8, pp. 46007–46016, 2020. View at: Publisher Site | Google Scholar
  8. E. Cambria, “Affective computing and sentiment analysis,” IEEE Intelligent Systems, vol. 31, no. 2, pp. 102–107, 2016. View at: Publisher Site | Google Scholar
  9. M. S. Rani and S. Sumathy, “Perspectives of the performance metrics in lexicon and hybrid based approaches: a review,” International Journal of Engineering and Technology, vol. 6, no. 4, pp. 108–115, 2017. View at: Publisher Site | Google Scholar
  10. E. Cambria, S. Poria, R. Bajpai, and B. Schuller, “Senticnet 4: A semantic resource for sentiment analysis based on conceptual primitives,” in Proceedings of COLING 2016, the 26th international conference on computational linguistics: Technical papers, pp. 2666–2677, Osaka, Japan, 2016. View at: Google Scholar
  11. A. Balahur, J. M. Hermida, and A. Montoyo, “Detecting implicit expressions of emotion in text: a comparative analysis,” Decision Support Systems, vol. 53, no. 4, pp. 742–753, 2012. View at: Publisher Site | Google Scholar
  12. J. A. Dominguez-Jimenez, K. C. Campo-Landines, J. Martinez-Santos, E. J. Delahoz, and S. Contreras-Ortiz, “A machine learning model for emotion recognition from physiological signals,” Biomedical Signal Processing and Control, vol. 55, article 101646, 2020. View at: Publisher Site | Google Scholar
  13. S. Sun, C. Luo, and J. Chen, “A review of natural language processing techniques for opinion mining systems,” Information Fusion, vol. 36, pp. 10–25, 2017. View at: Publisher Site | Google Scholar
  14. E. Cambria and A. Hussain, Sentic computing: Sentic computing: a common sense-based framework for concept-level sentiment analysis, Springer, 2015.
  15. M. Araujo, P. Goncalves, M. Cha, and F. Benevenuto, “ifeel: a system that compares and combines sentiment analysis methods,” in Proceedings of the 23rd International Conference on World Wide Web, ACM, pp. 75–78, Seoul Korea, 2014. View at: Publisher Site | Google Scholar
  16. S. Koelstra, C. Muhl, M. Soleymani et al., “Deap: a database for emotion analysis; using physiological signals,” IEEE Transactions on Affective Computing, vol. 3, no. 1, pp. 18–31, 2012. View at: Publisher Site | Google Scholar
  17. S. Katsigiannis and N. Ramzan, “Dreamer: a database for emotion recognition through eeg and ecg signals from wireless low-cost off-the-shelf devices,” IEEE Journal of Biomedical and Health Informatics, vol. 22, no. 1, pp. 98–107, 2018. View at: Publisher Site | Google Scholar
  18. K. Tripathi, Important Physiological Signals in the body, 2019, January 2019, http://biomedikal.in/2011/05/important-physiological-signals-in-the-body/.
  19. T. Matlovic, P. Gaspar, R. Moro, J. Simko, and M. Bielikova, “Emotions detection using facial expressions recognition and eeg,” in 2016 11th international workshop on semantic and social media adaptation and personalization (SMAP), pp. 18–23, Thessaloniki, Greece, 2016. View at: Google Scholar
  20. E. Douglas-Cowie, N. Campbell, R. Cowie, and P. Roach, “Emotional speech: towards a new generation of databases,” Speech Communication, vol. 40, no. 1-2, pp. 33–60, 2003. View at: Publisher Site | Google Scholar
  21. P. Petta, C. Pelachaud, and R. Cowie, Emotion-Oriented Systems: The HUMAINE Handbook, Springer, 2011.
  22. G. McKeown, M. Valstar, R. Cowie, M. Pantic, and M. Schroder, “The semaine database: annotated multimodal records of emotionally colored conversations between a person and a limited agent,” IEEE Transactions on Affective Computing, vol. 3, no. 1, pp. 5–17, 2012. View at: Publisher Site | Google Scholar
  23. C. Busso, M. Bulut, C.-C. Lee et al., “Iemocap: interactive emotional dyadic motion capture database,” Language Resources and Evaluation, vol. 42, no. 4, pp. 335–359, 2008. View at: Publisher Site | Google Scholar
  24. O. Martin, I. Kotsia, B. Macq, and I. Pitas, “The enterface'05 audio-visual emotion database,” in 22nd International Conference on Data Engineering Workshops (ICDEW'06), p. 8, Atlanta, GA, USA, 2006. View at: Google Scholar
  25. T. B. Alakus, M. Gonen, and I. Turkoglu, “Database for an emotion recognition system based on eeg signals and various computer games–GAMEEMO,” Biomedical Signal Processing and Control, vol. 60, p. 101951, 2020. View at: Publisher Site | Google Scholar
  26. D. H. Hockenbury and S. E. Hockenbury, Discovering Psychology, Macmillan, 2010.
  27. I. B. Mauss and M. D. Robinson, “Measures of emotion: a review,” Cognition and Emotion, vol. 23, no. 2, pp. 209–237, 2009. View at: Publisher Site | Google Scholar
  28. E. Fox, Emotion Science Cognitive and Neuroscientific Approaches to Understanding Human Emotions, Palgrave Macmillan, 2008. View at: Publisher Site
  29. R. Plutchik, “The nature of emotions,” American Scientist, vol. 89, no. 4, pp. 344–350, 2001. View at: Publisher Site | Google Scholar
  30. P. J. Lang, “The emotion probe: studies of motivation and attention,” American Psychologist, vol. 50, no. 5, pp. 372–385, 1995. View at: Publisher Site | Google Scholar
  31. “Manuals,” 2016, February 2014 https://www.trans-cranial.com/manuals/. View at: Google Scholar
  32. “Clinically Validated tDCS Technology,” 2016, https://www.trans-cranial.com/. View at: Google Scholar
  33. J. A. M. Correa, M. K. Abadi, N. Sebe, and I. Patras, “Amigos: a dataset for affect, personality and mood research on individuals and groups,” IEEE Transactions on Affective Computing, 2018. View at: Publisher Site | Google Scholar
  34. S. Aksoy and R. M. Haralick, “Feature normalization and likelihood-based similarity measures for image retrieval,” Pattern Recognition Letters, vol. 22, no. 5, pp. 563–582, 2001. View at: Publisher Site | Google Scholar
  35. G.-B. Huang, H. Zhou, X. Ding, and R. Zhang, “Extreme learning machine for regression and multiclass classification,” IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 42, no. 2, pp. 513–529, 2012. View at: Publisher Site | Google Scholar
  36. Y. Wang, F. Cao, and Y. Yuan, “A study on effectiveness of extreme learning machine,” Neurocomputing, vol. 74, no. 16, pp. 2483–2490, 2011. View at: Publisher Site | Google Scholar
  37. A. Seal, D. Bhattacharjee, M. Nasipuri, C. Gonzalo-Martin, and E. Menasalvas, “À-trous wavelet transform-based hybrid image fusion for face recognition using region classifiers,” Expert Systems, vol. 35, no. 6, article e12307, 2018. View at: Publisher Site | Google Scholar
  38. K. K. Sharma and A. Seal, “Modeling uncertain data using Monte Carlo integration method for clustering,” Expert Systems with Applications, vol. 137, pp. 100–116, 2019. View at: Publisher Site | Google Scholar
  39. A. Karlekar and A. Seal, “Soynet: soybean leaf diseases classification,” Computers and Electronics in Agriculture, vol. 172, article 105342, 2020. View at: Publisher Site | Google Scholar

Copyright © 2020 Ayan Seal et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views2440
Downloads619
Citations

Related articles

Article of the Year Award: Outstanding research contributions of 2020, as selected by our Chief Editors. Read the winning articles.