- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

Computational Intelligence and Neuroscience

Volume 2011 (2011), Article ID 519868, 12 pages

http://dx.doi.org/10.1155/2011/519868

## Comparison of Classification Methods for P300 Brain-Computer Interface on Disabled Subjects

Laboratorium voor Neuro- en Psychofysiologie, K.U.Leuven, Campus Gasthuisberg, O&N 2, Bus 1021, Herestraat 49, B-3000 Leuven, Belgium

Received 14 March 2011; Revised 26 May 2011; Accepted 4 July 2011

Academic Editor: Laura Astolfi

Copyright © 2011 Nikolay V. Manyakov et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

We report on tests with a mind typing paradigm based on a P300 brain-computer interface (BCI) on a group of amyotrophic lateral sclerosis (ALS), middle cerebral artery (MCA) stroke, and subarachnoid hemorrhage (SAH) patients, suffering from motor and speech disabilities. We investigate the achieved typing accuracy given the individual patient's disorder, and how it correlates with the type of classifier used. We considered 7 types of classifiers, linear as well as nonlinear ones, and found that, overall, one type of linear classifier yielded a higher classification accuracy. In addition to the selection of the classifier, we also suggest and discuss a number of recommendations to be considered when building a P300-based typing system for disabled subjects.

#### 1. Introduction

Research on *brain-computer interfaces* (BCIs) has witnessed a tremendous development in recent years [1] that has even been covered in the popular media. Although a lot of research has been done on invasive BCIs, leading to brain implants decoding neural activity directly, which are primarily tested on animals, noninvasive BCIs, for example, based on *electroencephalograms* (EEG) recorded on the subject’s scalp, have recently enjoyed an increasing visibility since they do not require any surgical procedure, and can therefore be more easily tested on human subjects. Several noninvasive BCI paradigms have been described in the literature, but the one we concentrate on relies on *event-related potentials* (ERPs, a stereotyped electrophysiological response to an internal or external stimulus [2]).

One of the most explored ERP components is the P300. It can be detected while a subject is shown two types of events with one occurring much less frequently than the other (“rare event”). The rare event elicits an ERP consisting of an enhanced positive-going signal component with a latency of about 300 ms after stimulus onset [2]. In order to detect ERPs, single-trial recordings are usually not sufficient, and recordings over several trials need to be averaged: the recorded signal is a superposition of the activity related to the stimulus and all other ongoing brain activity together with noise. By averaging, the activity that is time locked to a known event (e.g., the onset of the attended stimulus) is extracted as an ERP, whereas the activity that is not related to the stimulus onset is expected to be averaged out. The stronger the ERP signal, the fewer trials are needed, and *vice versa*.

There has been a growing interest in the ERP detection problem, as witnessed by the increased availability of BCIs that rely on it. A notorious example is the P300 speller [3], with which subjects are able to type words on a computer screen. This application meets the BCI’s primary goal, namely, to improve the quality of life of neurologically impaired patients suffering from pathologies such as amyotrophic lateral sclerosis (ALS), brain stroke, brain/spinal cord injury, cerebral palsy, muscular dystrophy, and so forth. But, as it is mostly the case with BCI research, the P300 BCI has primarily been tested on *healthy* subjects. Only very few attempts have been made on *patients* [4–9]. Several of these tests on patients [4, 9] deal with P300-based online typing, however, since only very few patients were tested, it is still an open question for which patient categories the P300 speller is best suited.

In addition, the performances of different P300 classifiers were compared for healthy subjects only, and their outcomes were found to disagree to some extent. In [10], a comparison of several classifiers (Pearson’s correlation method, Fisher’s linear discriminant analysis (LDA), stepwise linear discriminant analysis (SWLDA), linear support-vector machine (SVM), and Gaussian kernel support vector machine (nSVM)) was performed on 8 healthy subjects. It was shown that SWLDA and LDA render the best overall performance. In [11], it was shown that, among linear SVM, Gaussian kernel SVM, multi-layer perceptron, Fisher LDA, and kernel Fisher Discriminant, the best performance was achieved with LDA. Based on these studies, albeit different sets of classifiers were used, one can conclude that linear classifiers work better than nonlinear ones, at least for the case of the P300 BCI on healthy subjects. This statement is also supported by other researchers (e.g*.*, in [12]).

In light of this, and since a classifier comparison has never been performed on patients, it remains an open question what is the best classifier in this case. This is indeed an important question since the P300 responses from healthy subjects and patients can be quite different [5]. Thus, the outcome of a comparison for healthy subjects might not be valid for patients.

In this paper, we report on tests performed on a group of (partially) disabled patients suffering from amyotrophic lateral sclerosis (ALS), middle cerebral artery (MCA) stroke, and subarachnoid hemorrhage (SAH). In addition to the classifiers mentioned above, we also add two more linear ones (i.e*.*, Bayesian linear discriminant analysis and a method based on feature extraction), since they have been used before in P300 BCIs [7, 13]. In summary, we compare a more extensive set of classifiers and perform our comparison on patients, instead of on healthy subjects, both of which distinguish our study from others.

#### 2. Methods

##### 2.1. EEG Data Acquisition

Our recordings were performed with a prototype of a* miniature* EEG recording device that * wirelessly* communicates with a USB stick receiver (Figures 1(a) and 1(c)). The prototype was developed by *imec *(http://www.imec.be/) and built around their ultra-low power 8-channel EEG amplifier chip [14]. The EEG data were recorded at a sampling frequency of 1000 Hz, which is fixed by the hardware. A laptop working under Windows XP SP3 with a bright 15′′ screen was used for the visual stimulation as well as for EEG data recording, processing and storing.

We used an electrode cap with large filling holes and sockets for active Ag/AgCl electrodes (ActiCap, Brain Products, Figure 1(d)). The eight electrodes were placed primarily on the parietal pole, namely at positions Cz, CPz, P1, Pz, P2, PO3, POz, and PO4, according to the international 10–10 system (Figure 1(b)). The reference and ground electrodes were placed on the left and right mastoids, respectively.

Each experiment started with a pause of approximately 90 s, which is required for the EEG amplifier to stabilize its internal filters. During this period, the EEG signals were not recorded. The data for typing each character (see Section 2.3 for details) were recorded in one session. As the duration of each session is known *a priori*, as well as the data transfer rate, it is easy to estimate the amount of data transmitted during a session. We used this estimate, increased by a margin, as the size of the serial port buffer. To make sure that the entire recording session for one character fits completely into the buffer, we cleared the buffer just before recording. This strategy allowed us to avoid broken/lost data frames, which might occur due to a buffer overflow. The EEG data frames were only in rare cases lost during wireless transmission: under normal experimental conditions, the data loss is negligible (<0.01%) and never more than a few consecutive samples, which could be (linearly) reconstructed from the successfully received ones. The amount of broken/lost frames can be precisely computed using the counter incorporated into each data frame.

##### 2.2. Data-Stimuli Synchronization

Unlike a conventional EEG system, the system we used does not have any external synchronization inputs. We used a synchronization scheme based on the high-precision timestamps of the stimulus onsets (during stimulation). The timestamps are obtained using the high-resolution system performance counter (via QueryPerformanceCounter system call), which allows to achieve microsecond resolution. We save the timestamps of the EEG acquisition session start and end, as well as the timestamps of the stimulus onsets and offsets. Due to the fact that the EEG signal has a constant sampling rate, and assuming a constant (virtual) serial port latency, the precise mapping between the timestamps and the corresponding EEG data samples is straightforward. We used this mapping for partitioning the EEG signal into signal segments, for further processing. To eliminate the unwanted load of the computer system, all recordings were done on a PC working in a special (“experimental”) hardware profile, which has a minimal set of running services and devices. Additionally, we have raised the priority of the application, responsible for the visual stimulation and EEG data acquisition/processing, to the “high” level.

##### 2.3. Experiment Design

Twelve subjects, naïve to BCI applications, participated in the experiments (ten male and two female, aged 37–66 with an average age of 51.25). The subjects were suffering from different types of brain disorders. The experimental protocol was approved by the ethical committee. After the recordings were made, four subjects were excluded from the classifier comparison, since their performance was close to chance level, possibly due to the nature of their brain disorder, or because they did not properly understand the experiment or were too tired to perform the task. The information about the patients, of which the recordings were further considered, that is, their diagnosis, age, and gender, is presented in Table 1.

We have used the same visual stimulus paradigm as in the first P300-based speller, introduced by Farwell and Donchin in [3]: a matrix of characters. The only difference is with the character set: in our case, it is the usual set of 26 Latin characters, eight digits, but with two special characters (“_” instead of *space* and “” as the *end of input* indicator). Additionally, for some subjects, the Cyrillic alphabet was used. Each experiment was composed of one training and several testing stages. During both stages, columns and rows of the matrix were intensified (see Figure 2) in a random manner. The duration of the intensification was fixed to 100 ms, followed by 100 ms of no intensification. Each column and each row flashed only once during one trial, so each trial consisted of 12 stimulus presentations.

During the training stage, 11 characters, taken from the typing matrix, were presented to the subject. For each character, 10 intensifications for each row/column were performed. The subjects were asked to mentally count the number of intensifications of the intended character. The counting was used only to ensure that the participants paid attention.

The recorded data was filtered (in the 0.5–15 Hz frequency band with a fourth-order zero-phase digital Butterworth filter) and cut into signal segments. Each of these segments consisted of 1000 ms of recording, starting from the stimuli onsets. Then, they were downsampled, by retaining every 25th sample, and assigned to one of two possible groups: *target* and *nontarget*, according to the stimuli that they were locked to. For training the classifier, we constructed a set of 1000 target, and the same amount of nontarget averaged brain responses, where the averages were taken based on randomly selected responses from the corresponding groups in the training set. The number was equal to the number of intensification sequences (trials), for each stimulus, during the testing stage.

Signal amplitudes at specific time instants in the interval 100–750 ms after stimuli onset, of the downsampled EEG signal, were taken as features. All these features were normalized to their *Z* score through the estimation of , where is the EEG amplitude of th channel (electrode) at time , after stimulus onset, the average of , and the standard deviation for all training examples of both the target and nontarget recordings of the training set. When combining all those features, we obtained a feature vector , which was used as input to either the linear classifier (see further) or the nonlinear one, . Since we use scores as features, and since we use a balanced training set (equal numbers of target and nontarget responses), the parameter should be close to zero. After substituting the feature vector into the above-mentioned equation, we obtain a “distance” from the point in feature space to the boundary (hyperplane in the linear case), separating the target from the nontarget class, with the sign indicating to which class the point belongs.

After training the classifier, each subject performed several *online* test sessions during which (s)he was asked to *mind-type* a few words. In the case of a mistyping, (s)he was instructed to type further without trying to correct the mistake (“backspace” was not allowed). The typing performance (ratio of correctly typed characters) was used for estimating the classification accuracy. For these online test sessions, we considered a linear SVM classifier trained on data averaged over 15 trials. Thus, each subject attempted to type characters based on 15 row/column intensifications. About 36 characters were typed by each subject. This number slightly varied between subjects, since some subjects chose the characters to spell themselves (free spelling). During typing, the EEG data was stored for further (*offline*) analysis based on a smaller amount of trials (in this case we used all -combination of 15 trials for each typed character, for assessing the accuracy).

The testing stage differs from the training stage by the way the signal segments were grouped. During training, the system “knows” exactly which one of 36 possible characters is attended by the subject at any moment of time (copy spelling). Based on this information, the collected signal segments can be grouped into only two categories: target (attended) and nontarget (not attended). However, during testing, the system does not know which character is attended by the subject, and the only meaningful way of grouping is by stimulus type (which in the proposed paradigm can be one of 12 types: 6 rows and 6 columns). Thus, during the testing stage, for each trial, we had 12 segments (from all 12 types) of 1000 ms EEG data recorded from each electrode. The averaged EEG response for each electrode was determined for each stimulus type. The selected features of the averaged data were then fed into the classifier (see Section 3). As a result, the classifier produces 12 (for each row/column) values which describe the distance to the class boundary in the feature space, together with the sign. The row index and the column index of the classified character were calculated as The character at the intersection of the row and column in the matrix was then taken as the result of the classification and presented, as a feedback, to the subject, in online mode.

#### 3. Classification Methods

##### 3.1. Fisher’s Linear Discriminant Analysis

Fisher's linear discriminant analysis (LDA) is one of the most widely used classifiers in P300 BCI systems [10, 15]. It was reported to even outperform other classifiers [11]. Its main idea is to find a projection from the -dimensional feature space onto a one-dimensional space for which the ratio of the variance between the two classes (target and nontarget) versus the variance within the classes is maximal. This “optimal” projection is estimated as , where and define the covariances and the means of the two classes (target and nontarget) that need to be separated.

##### 3.2. Stepwise Linear Discriminant Analysis

Stepwise linear discriminant analysis (SWLDA) has been used in patient studies of the P300 BCI speller [4, 5]. It can be considered as an extension of the LDA with an incorporated filter feature selection. SWLDA adds and removes terms from a linear discriminant model, based on their statistical significance in regression, thus, producing model that is adjustable to the training data. It was shown that SWLDA performs equally well or even better than several other classification methods in P300 BCIs [10]. For our comparison, we have used the same procedure as in [10] (in the forward step, the entrance tolerance ; in the backward step, the exit tolerance ). The process was iterated until convergence, or until it reached a predefined number of 60 features.

##### 3.3. Bayesian Linear Discriminant Analysis

Bayesian linear discriminant analysis (BLDA) has been used in P300 BCI patient studies [7]. It is based on a probabilistic regression network. Suppose that the targets (in the case of a classification problem these are and −1) are linearly dependent on the observed features with an additive Gaussian noise term : . Assuming further an independent generation of the examples from a data set, the likelihood of all data is . Additionally to this, we have to introduce a prior distribution over all weights as a zero-mean Gaussian Using Bayes’s rule, we can define the posterior distribution which is a Gaussian with mean and covariance matrix , where is the identity matrix, a matrix with each row corresponding to a training example in feature space, and a column vector of true labels (classification) for all corresponding training examples. As a result, our hyperplane will have the form . This solution is equivalent to a penalized least square estimate [16]. Regression parameters ( and ) are tuned with an automatic, iterative procedure [7].

##### 3.4. Linear Support Vector Machine

In P300 BCI research, the linear support vector machine (SVM) is regarded as one of the more accurate classifiers [10, 17]. The principal idea of a linear SVM is to find the separating hyperplane, between two classes, so that the distance between the hyperplane and the closest points from both classes is maximal. In other words, we need to maximize the margin between the two classes [18]. Since it is not always the case that the two classes are linearly separable, the linear SVM idea was also generalized to the case where the data points are allowed to fall within the margin (and even are on the wrong side of the decision boundary) by adding a regularization term. For our analysis, we used the method based on linear least squares SVM [19] to solve the minimization problem with respect to , , where corresponds to the training points in the feature space, and is the associated output ( for the responses to the target stimulus and −1 for the nontarget stimulus). The regularization parameter is estimated through a line search on cross-validation results.

##### 3.5. Nonlinear Support Vector Machine

Here, we used a support vector machine with the Gaussian radial-basis function , , as a kernel. In our experiment, we opted for the SVMlight package [20]. The SVM’s outcome, for a new sample, is a value for , where are the support vectors chosen from the training set with known class labels , and where are Lagrange multipliers. The sign of estimates the class the sample belongs to. For our nSVM classifier, a search through pairs (where is the regularization parameter and the kernel parameter) was performed using a 5-fold cross-validation on the grid .

##### 3.6. Method Based on Feature Extraction

Another linear classifier used in P300 BCI research [13] relies on the one-dimensional version of the linear *feature extraction* (FE) approach proposed by Leiva-Murillo and Artès-Rodriguez in [21]. The method searches for the “optimal” subspace maximizing (an estimate of) the mutual information between the set of projections and the set of corresponding labels . According to [21], the mutual information between the set of projections and the set of corresponding labels can be estimated as , with the number of classes, the projection of the th class' data points onto the direction , the standard deviation, and the negentropy, estimated using Hyvärinen’s robust estimator [22].

##### 3.7. Artificial Neural Network

For comparison's sake, we also consider a multilayer feed-forward neural network (NN) with a single hidden layer and with sigmoidal activation functions, which is proved to be a universal approximator [23]. Thus, our classifier has the form where is the number of neurons in the hidden layer, with sigmoidal activation functions , the number of observed features, and sets of thresholds and weight coefficients, respectively. The latter were optimized using a training procedure based on the Levenberg-Marquardt back propagation method, where the desired outcome of the neural network was set to or −1 (target or nontarget), depending on the class the individual training example belongs to. Since such a network has parameters to be trained, it can easily overfit the training data in the case of a large number of features (), and a large number of hidden layer neurons (). To avoid this, we performed a 5-fold cross-validation with a line search for the number of hidden neurons . The network with the best was further retrained on the whole training set.

#### 4. Results

The data was recorded during the * online* typing of words/characters (in copy spell and in free spell mode). In order to assess the classification performance of all classifiers considered, we opted for an *offline* analysis, in which case we also evaluated the performance for a smaller amount of intensification sequences . This became possible since our online spelling was performed with 15 intensifications of each row and column for any character to be typed. This also allowed us to construct a larger amount of test data for . This was done by taking combinations of elements from the available 15 responses for each row and column.

The performance results are shown in Figure 3 for each individual patient, and the averaged performance result in Figure 4, averaged over all subjects. In order to verify the statistical significance of the comparison, we used a repeated-measures two-way ANOVA (with “method” and “intensification sequences” as factors) with Greenhouse-Geisser correction ( for factor “method”) and with *post hoc* multiple comparison based on Turkey LSD test for pairs of all methods. We found that the accuracy of a BLDA in general is significantly () better than that of any other classifier except the Gaussian kernel SVM (nSVM versus BLDA has ), since the later, for some subjects, and for some numbers of intensifications , yielded on average better results. Both the linear and nonlinear SVM’s (for which the results do not show any significant difference) were second best. As for SWLDA and LDA, which ranked third, SWLDA performs slightly better, but not in a significant way. The worst results are obtained for the feature extraction (FE) method and the multilayer feed-forward neural network (NN).

We have also analyzed the distribution of the erroneously typed characters (see Figure 5). We have found that, for all classifiers, the misclassifications mostly occur for either a row or a column in close proximity to the ones of the intended characters (represented at the center of the plot). To investigate any possible differences in the error distributions for each of the considered classifiers, we computed the horizontal (for the columns) and the vertical (for the rows) standard deviations (std) between the typed and the intended characters, and plot this as a function of the number of intensifications (Figure 6). The BLDA classifier for the case of the rows and BLDA together with nSVM for the case of the columns yield, in general, the smallest std, suggesting that those classifiers lead to less wrong answers. In order to verify the statistical significance of the comparison, we used a repeated-measures three-way ANOVA for std using the following factor levels: “method” (with further * post hoc* multiple comparison of all pairwise combinations of classifiers), “direction” (with two levels for this factor, corresponding to rows and columns), and “intensification sequences” (15 levels). We found that the distribution of mistakes around the intended character, based on BLDA, is, in general, significantly ( for factor “method”) smaller than for any other classifier, except for nSVM (nSVM versus BLDA has ). This suggests that the BLDA, in general, not only yields a better accuracy, but also leads to a smaller divergence in mistakes. We also observe that the vertical standard deviation is in general smaller than the horizontal one ( for factor “direction”), particularly for the most accurate classifiers and, especially, after more than 5-6 intensification sequences. For example, for BLDA (fixing this level of factor “method” in previous model), this difference is significant with .

#### 5. Discussion

Our comparison indicates that, in general, nonlinear classifiers perform worse or equal to linear ones. This is in accordance with other studies [10–12], which were performed on healthy subjects. This could be due to the tendency of nonlinear classifiers to overfit the training data, leading to an inferior generalization performance. It is mostly relevant for the multilayer feed-forward neural network, since the kernel SVM is known to properly deal with high dimensional data and small training sets [18]. In our study, the Gaussian kernel SVM generates a result that is not significantly different from its linear counterpart, but at the expense of an exhaustive grid search. From this, we recommend a linear classifier for a P300 spelling systems for patients, also since, to support its *online* applicability, we have to minimize the classifier’s training time.

Among all classifiers the Bayesian linear discriminant analysis (BLDA) yields superior results, with the SVM as the second best, at least for the group of patients considered in our comparison. While a SVM is constructed so as to maximize a margin between the two classes, the BLDA tries to maximize the probability of having training data with the correct class labels. Since both classifiers depend on some regularization parameters, their optimal choice increases the generalization accuracy. This optimization enables us to achieve better results for the P300 speller based on SVM and BLDA. While in SVM, the parameter optimization is done with a search through a discrete set of parameters, in the framework of a cross-validation (thus, depending on the search algorithm, and the resolution of the discretization), BLDA includes a self-adjustment of its parameters via an automatic, iterative procedure. On the other hand, BLDA relies on assumed distributions of the classification errors and of the used parameters.

From the obtained classification results, we observe that different classifiers lead to different accuracies. On the one hand, this shows the necessity to properly choose the classifier for the intended P300 BCI application. But on the other hand, this diversity in results could be turned into a benefit by combining different classifiers in a co-training approach [15], to improve the classification performance.

For the validation of the performance of the classifiers and their comparison, we used as features the amplitudes of the filtered EEG signals from different electrodes. This led to satisfactory results for healthy subjects (see, *e.g.*, [17]). Nevertheless, the accuracy could potentially be improved by adding other features such as time-frequency ones, from a wavelet transform [24], synchrony between EEG channels [25], and the direction and speed of propagating waves [26].

In our experiments, we used electrodes placed at positions Cz, CPz, P1, Pz, P2, PO3, POz, and PO4, which include the parietal ones for which the P300 component is known to be most prominent, but we also added more posterior positions, as suggested in [7, 27–29] where it was shown that the decoding accuracy increases due to the negative-going component, appearing over the posterior areas, prior to the P300 component. To incorporate this additional early information into the decoding process, we used the interval starting 100 ms after stimulus onset. The negative-going component, called N2 in [30], was shown by these authors to be important for the P300 speller, even if the subject only covertly attended the intended target. Thus, for patients, when experiencing problems with eye gazing, the early negative component recorded over the posterior positions seems to be beneficial.

To validate the added value of the different ERP components into the decoding performance, we estimated the classification accuracy in the P300 speller with 15 intensification sequences and the BLDA classifier, for each patient separately, and for the features taken from 50 ms time intervals (the centers of these intervals were spaced by 25 ms). The classification results are shown in Figure 7, and the averaged ERP waveforms in Figure 8, for electrode POz. The results suggest that the early ERP components should, for some of our patients, also be considered as features for decoding.

The analysis of the distribution of the mistyped characters (Figure 5) suggests that mistakes mostly occur due to a wrongly selected row or column in the typing matrix. Furthermore, we found that the incorrectly typed characters are mostly close to the intended ones. This could, probably, be due to the fact that the subject sometimes gets distracted by the flashing of a column or row adjacent to one containing the intended character. Or, it could be that the intensification of the row/column containing the intended character is immediately preceded or followed by an intensification of an adjacent row/column, leading to a decreased P300 response. As a recommendation, one should try to avoid the consecutive intensifications of adjacent rows/columns. But this is hard to achieve in a row/column paradigm, since in a free spelling mode we do not know *a priori* the character that the subject wants to communicate. Additionally to this, based on the fact that mistakes mostly occur along the row or column containing the desired character, we can try to use some smart scrambling of the intensifications where, instead of a whole row or column, constellations of individual characters, spread over the entire matrix, are intensified. The design of the proper stimulation paradigm as in, for example, [31] is the subject of further research.

Another way to improve the typing performance is by incorporating the detection of the Error Potential (ErrP) [32, 33] into the P300 speller paradigm. The ErrP is evoked when the subject perceives a wrong outcome of the BCI system. When the ErrP is detected, we can take the second most likely character (e.g., the row or the column with the second largest distance to the classification boundary) for correcting the classifier’s outcome. Since mistakes are expected to occur in a row or column adjacent to that of the desired character in the matrix (see Figure 5), we can also apply weights to the previous distances (e.g., by inversely relating them to the distance, in the matrix, to the mistyped character).

The typing accuracies achieved by our patients revealed a large variability. While subjects 2 and 8 could achieve an almost perfect typing performance for already row/column intensifications, subjects 4 and 7 achieved the worst accuracy (around 50% after intensifications, with a chance level of ). As can be seen from Table 1, the latter subjects suffered from some form of motor aphasia (as was also the case with three of the four subjects excluded from the classifier comparison study because of bad classification performance (see Section 2.3)). Motor aphasia is known to deteriorate the visual verbal P300 latency more than the visual nonverbal one [34], possibly explaining the inferior performance achieved with these patients. The effect on the P300 speller should be examined further in a study specifically designed for motor aphasia patients.

#### 6. Conclusions

We have compared five linear and two nonlinear classifiers in a P300 BCI speller tested on stroke and ALS patients. We have found that the BLDA classifier performs the best, followed by the (non)linear SVM. These results could be helpful to decide what classifier to use for stroke and ALS patients. Finally, we also listed and discussed a number of recommendations for adjusting the P300 speller paradigm to stroke and ALS patients.

#### Acknowledgments

NVM is supported by the Flemish Regional Ministry of Education (Belgium) (GOA 10/019). NC is supported by the European Commission (IST-2007-217077). AC is supported by a specialization Grant from the Agentschap voor Innovatie door Wetenschap en Technologie (IWT, Flemish Agency for Innovation through Science and Technology). MMVH is supported by research Grants received from the Financing program (PFV/10/008) and the CREA Financing program (CREA/07/027) of the K.U.Leuven, the Belgian Fund for Scientific Research—Flanders (G.0588.09), the Interuniversity Attraction Poles Programme—Belgian Science Policy (IUAP P6/29), the Flemish Regional Ministry of Education (Belgium) (GOA 10/019), and the European Commission (IST-2007-217077), and by the SWIFT prize of the King Baudouin Foundation of Belgium. The authors wish to thank Valiantsin Raduta and Yauheni Raduta of the Neurology Department of the Brest Regional Hospital (Brest, Belarus) for selecting the patients and their assistance in the recordings. The authors are also grateful to Refet Firat Yazicioglu, Tom Torfs, and Chris Van Hoof from imec, Leuven, for providing us with the wireless EEG system. Finally, they would like to thank Prof. Philip Van Damme from the Experimental Neurology Department, Katholieke Universiteit Leuven, for his assistance in translating the patient diagnoses from Russian.

#### References

- P. Sajda, K. R. Müller, and K. V. Shenoy, “Brain-computer interfaces,”
*IEEE Signal Processing Magazine*, vol. 25, no. 1, pp. 16–17, 2008. View at Publisher · View at Google Scholar · View at Scopus - S. Luck,
*An Introduction to the Event-Related Potential Technique*, MIT Press, Cambridge, Mass, USA, 2005. - L. A. Farwell and E. Donchin, “Talking off the top of your head: toward a mental prosthesis utilizing event-related brain potentials,”
*Electroencephalography and Clinical Neurophysiology*, vol. 70, no. 6, pp. 510–523, 1988. View at Google Scholar · View at Scopus - F. Nijboer, E. W. Sellers, J. Mellinger et al., “A P300-based brain-computer interface for people with amyotrophic lateral sclerosis,”
*Clinical Neurophysiology*, vol. 119, no. 8, pp. 1909–1916, 2008. View at Publisher · View at Google Scholar · View at Scopus - E. W. Sellers and E. Donchin, “A P300-based brain-computer interface: initial tests by ALS patients,”
*Clinical Neurophysiology*, vol. 117, no. 3, pp. 538–548, 2006. View at Publisher · View at Google Scholar · View at Scopus - F. Piccione, F. Giorgi, P. Tonin et al., “P300-based brain computer interface: reliability and performance in healthy and paralysed participants,”
*Clinical Neurophysiology*, vol. 117, no. 3, pp. 531–537, 2006. View at Publisher · View at Google Scholar · View at Scopus - U. Hoffmann, J. M. Vesin, T. Ebrahimi, and K. Diserens, “An efficient P300-based brain-computer interface for disabled subjects,”
*Journal of Neuroscience Methods*, vol. 167, no. 1, pp. 115–125, 2008. View at Publisher · View at Google Scholar · View at Scopus - S. Silvoni, C. Volpato, M. Cavinato, et al., “P300-based brain-computer interface communication: evaluation and follow-up in amyotrophic lateral sclerosis,”
*Frontiers in Neuroscience*, vol. 3, no. 60, pp. 1–12, 2009. View at Google Scholar - E. W. Sellers, T. M. Vaughan, and J. R. Wolpaw, “A brain-computer interface for long-term independent home use,”
*Amyotrophic Lateral Sclerosis*, vol. 11, no. 5, pp. 449–455, 2010. View at Publisher · View at Google Scholar · View at Scopus - D. J. Krusienski, E. W. Sellers, F. Cabestaing et al., “A comparison of classification techniques for the P300 Speller,”
*Journal of Neural Engineering*, vol. 3, no. 4, pp. 299–305, 2006. View at Google Scholar · View at Scopus - H. Mirghasemi, R. Fazel-Rezai, and M. B. Shamsollahi, “Analysis of P300 classifiers in brain computer interface speller,” in
*Proceedings of the 28th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS '06)*, pp. 6205–6208, September 2006. View at Publisher · View at Google Scholar · View at Scopus - F. Lotte, M. Congedo, A. Lécuyer, F. Lamarche, and B. Arnaldi, “A review of classification algorithms for EEG-based brain-computer interfaces,”
*Journal of Neural Engineering*, vol. 4, no. 2, pp. R1–R13, 2007. View at Google Scholar · View at Scopus - N. Chumerin, N. V. Manyakov, A. Combaz et al., “P300 detection based on feature extraction in on-line brain-computer interface,”
*Lecture Notes in Computer Science*, vol. 5803, pp. 339–346, 2009. View at Publisher · View at Google Scholar · View at Scopus - R. F. Yazicioglu, P. Merken, R. Puers, and C. Van Hoof, “Low-power low-noise 8-channel EEG front-end ASIC for ambulatory acquisition systems,” in
*Proceedings of the 32nd European Solid-State Circuits Conference (ESSCIRC '06)*, pp. 247–250, September 2006. View at Publisher · View at Google Scholar · View at Scopus - R. C. Panicker, S. Puthusserypady, and Y. Sun, “Adaptation in P300 braincomputer interfaces: a two-classifier cotraining approach,”
*IEEE Transactions on Biomedical Engineering*, vol. 57, no. 12, pp. 2927–2935, 2010. View at Publisher · View at Google Scholar · View at Scopus - M. E. Tipping, “Bayesian inference: an introduction to principles and practice in machine learning,” in
*Advanced Lectures on Machine Learning*, O. Bousquet, U. von Luxburg, and G. Rätsch, Eds., pp. 41–62, Springer, New York, NY, USA, 2004. View at Google Scholar - M. Thulasidas, C. Guan, and J. Wu, “Robust classification of EEG signal for brain-computer interface,”
*IEEE Transactions on Neural Systems and Rehabilitation Engineering*, vol. 14, no. 1, pp. 24–29, 2006. View at Publisher · View at Google Scholar · View at Scopus - V. Vapnik,
*The Nature of Statistical Learning Theory*, Springer, New York, NY, USA, 1995. - J. Suykens, T. Van Gestel, J. De Brabanter, B. De Moor, and J. Vanderwalle,
*Least Square Support Vector Machines*, World Scientific, Singapore, 2002. - T. Joachims, “Making large-scale SVM learning practical,” in
*Advance in Kernel Methods—Support Vector Learning*, B. Schölkopf, C. Burges, and A. Smola, Eds., pp. 169–184, MIT Press, Cambridge, Mass, USA, 1999. View at Google Scholar - J. M. Leiva-Murillo and A. Artés-Rodríguez, “Maximization of mutual information for supervised linear feature extraction,”
*IEEE Transactions on Neural Networks*, vol. 18, no. 5, pp. 1433–1441, 2007. View at Publisher · View at Google Scholar - A. Hyvärinen, “New approximations of differential entropy for independent component analysis and projection pursuit,” in
*Proceedings of the Conference on Advances in Neural Information Processing Systems*, pp. 273–279, MIT Press, Cambridge, Mass, USA, 1998. - G. Cybenko, “Approximation by superpositions of a sigmoidal function,”
*Mathematics of Control, Signals, and Systems*, vol. 2, no. 4, pp. 303–314, 1989. View at Publisher · View at Google Scholar · View at Scopus - V. Bostanov and B. Kotchoubey, “The t-CWT: a new ERP detection and quantification method based on the continuous wavelet transform and Student's t-statistics,”
*Clinical Neurophysiology*, vol. 117, no. 12, pp. 2627–2644, 2006. View at Publisher · View at Google Scholar · View at Scopus - E. Gysels and P. Celka, “Phase synchronization for the recognition of mental tasks in a brain-computer interface,”
*IEEE Transactions on Neural Systems and Rehabilitation Engineering*, vol. 12, no. 4, pp. 406–415, 2004. View at Publisher · View at Google Scholar · View at Scopus - N. V. Manyakov, R. Vogels, and M. M. Van Hulle, “Decoding stimulus-reward pairing from local field potentials recorded from monkey visual cortex,”
*IEEE Transactions on Neural Networks*, vol. 21, no. 12, pp. 1892–1902, 2010. View at Publisher · View at Google Scholar · View at Scopus - D. J. Krusienski, E. W. Sellers, D. J. McFarland, T. M. Vaughan, and J. R. Wolpaw, “Toward enhanced P300 speller performance,”
*Journal of Neuroscience Methods*, vol. 167, no. 1, pp. 15–21, 2008. View at Publisher · View at Google Scholar · View at Scopus - E. Sellers, D. Krusienski, D. McFarland, and J. Wolpaw, “Non-invasive brain-computer interface research at the Wadsworth Center,” in
*Toward Brain-Computer Interfacing*, G. Dornhege, J. Millán, T. Hinterberger, D. McFarland, and K.-R. Müller, Eds., pp. 31–42, MIT Press, Cambridge, Mass, USA, 2007. View at Google Scholar - S. L. Shishkin, I. P. Ganin, I. A. Basyul, A. Y. Zhigalov, and A. Ya. Kaplan, “N1 wave in the P300 BCI is not sensitive to the physical characteristics of stimuli,”
*Journal of Integrative Neuroscience*, vol. 8, no. 4, pp. 471–485, 2009. View at Publisher · View at Google Scholar - M. S. Treder and B. Blankertz, “(C)overt attention and visual speller design in an ERP-based brain-computer interface,”
*Behavioral and Brain Functions*, vol. 6, article 28, 2010. View at Publisher · View at Google Scholar · View at Scopus - G. Townsend, B. K. LaPallo, C. B. Boulay et al., “A novel P300-based brain-computer interface stimulus presentation paradigm: moving beyond rows and columns,”
*Clinical Neurophysiology*, vol. 121, no. 7, pp. 1109–1120, 2010. View at Publisher · View at Google Scholar · View at Scopus - B. Dal Seno, M. Matteucci, and L. Mainardi, “Online detection of P300 and error potentials in a BCI speller,”
*Computational Intelligence and Neuroscience*, vol. 2010, Article ID 307254, 5 pages, 2010. View at Publisher · View at Google Scholar · View at Scopus - A. Combaz, N. Chumerin, N. V. Manyakov, A. Robben, J. A. K. Suykens, and M. M. Van Hulle, “Error-related potential recorded by EEG in the context of a P300 mind speller brain-computer interface,” in
*Proceedings of the 20th IEEE International Workshop on Machine Learning for Signal Processing (MLSP '10)*, pp. 65–70, September 2010. View at Publisher · View at Google Scholar · View at Scopus - R. Neshige, N. Murayama, W. Izumi, T. Igasaki, and K. Takano, “Non-verbal and verbal P300 of auditory and visual stimuli in dementia, dysarthria and aphasia,”
*Japanese Journal of Rehabilitation Medicine*, vol. 35, pp. 164–169, 1998. View at Google Scholar