Computational and Mathematical Methods in Medicine

Computational and Mathematical Methods in Medicine / 2013 / Article

Research Article | Open Access

Volume 2013 |Article ID 297860 | 15 pages |

Estimation of Phoneme-Specific HMM Topologies for the Automatic Recognition of Dysarthric Speech

Academic Editor: Volkhard Helms
Received31 May 2013
Revised16 Aug 2013
Accepted25 Aug 2013
Published08 Oct 2013


Dysarthria is a frequently occurring motor speech disorder which can be caused by neurological trauma, cerebral palsy, or degenerative neurological diseases. Because dysarthria affects phonation, articulation, and prosody, spoken communication of dysarthric speakers gets seriously restricted, affecting their quality of life and confidence. Assistive technology has led to the development of speech applications to improve the spoken communication of dysarthric speakers. In this field, this paper presents an approach to improve the accuracy of HMM-based speech recognition systems. Because phonatory dysfunction is a main characteristic of dysarthric speech, the phonemes of a dysarthric speaker are affected at different levels. Thus, the approach consists in finding the most suitable type of HMM topology (Bakis, Ergodic) for each phoneme in the speaker’s phonetic repertoire. The topology is further refined with a suitable number of states and Gaussian mixture components for acoustic modelling. This represents a difference when compared with studies where a single topology is assumed for all phonemes. Finding the suitable parameters (topology and mixtures components) is performed with a Genetic Algorithm (GA). Experiments with a well-known dysarthric speech database showed statistically significant improvements of the proposed approach when compared with the single topology approach, even for speakers with severe dysarthria.

1. Introduction

The term dysarthria was initially defined as “a collective name for a group of speech disorders resulting from disturbances in muscular control over the speech mechanism due to damage of the central or peripheral nervous system” [1, 2]. More recently, dysarthria is described as an impairment in one or more of the processes involved in speech production: respiration, phonation (control of the vocal folds for appropriate voice quality and valving of the airway), resonance (ability to alter the vocal tract, and resonating spaces for correct speech sounds), articulation, and prosody (appropriate emphasis and inflection) [3]. People affected by this condition often present significant deficiencies in oral communication and reduced intelligibility due to the characteristic paralysis, weakness, and incoordination of the speech musculature [4]. This affects all aspects in the life of dysarthric people, from social interaction to academic performance and vocational placement [5].

The damage of the nervous system that leads to dysarthria can be caused by congenital disorders (e.g., cerebral palsy), cerebrovascular accident (CVA), traumatic brain injury (TBI), or degenerative neurological disease such as Parkinson’s or Alzheimer’s disease. Some of these conditions as cerebral palsy and TBI produce nonprogressive dysarthria while Parkinson’s and Alzheimer’s disease produce a degenerative dysarthria that degrades speech over time [3].

The affected muscles in dysarthria may include the lungs, larynx, oropharynx and nasopharynx, soft palate, and articulators (lips, tongue, teeth, and jaw) [6]. The degree to which these muscle groups are compromised determines the particular pattern of speech impairment and the type (faccid, spastic, ataxic, hypokinetic, hyperkinetic) and level (mild, moderate, severe) of dysarthria. In Figure 1 a general overview of the speech impairments associated to each type of dysarthria as reported in [1] is presented, noting that most of the abnormalities are related to the processes of phonation, prosody, and articulation. Thus, dysarthric speech may include imprecise consonants and distorted vowels (segmental deficits), irregular articulatory breakdowns, excessive or equal stress to all syllables, and a slow rate of speech with a phonatory-prosodic insufficiency described as harsh, monotonous, and monoloudness [3]. Typical symptoms also include strained phonation, imprecise placement of the articulators, incomplete consonant closure resulting in sonorant implementation of many stops and fricatives, and reduced voice onset time (VOT) distinctions between voiced and unvoiced stops [4]. Because the range of speech impairments or abnormalities in dysarthric speech is extensive, assessment commonly requires different tests.

1.1. Effect of Dysarthria on Vowels, Consonants, and Prosody

Vowel distortion is common because dysarthric speakers produce individual movements or changes in overall vocal tract shape with reduced displacements and velocities [4]. This leads to the following specific abnormalities: large deviations and centralization in formant frequencies, change in vowel space area, shallower formant slopes, and greater interspeaker formant transition variability [4]. However, for dysarthric speakers, vowels are physically easy to produce since they do not require dynamic movement of the articulatory muscles.

In contrast to vowels, consonants require fine motor control of the articulators. Thus, these are more affected and difficult to produce in the presence of dysarthria [7]. This is particularly more evident for the consonants that involve some kind of frication event: a burst or transient noise (stops), a brief noise interval (affricates), or a longer noise interval (fricatives).

A metric used to measure the impairment of consonant production is the VOT, which is defined as the length of time that passes between the release of a stop consonant and the onset of voicing, the vibration of the vocal folds. Dysarthria is characterized by significant variability in VOT. For example, in [8] people with spastic dysarthria produced consonants with shorter VOT when compared with people with normal speech. In contrast, in [9, 10] long VOT durations were observed in people with spastic dysarthria. In [11] high VOT variability was observed among people with ataxic dysarthria.

About prosody, decreased range of pitch and loudness have been noted as prosodic characteristics of dysarthria [1] where structures of the language such as stress, rhythm, and intonation are affected [12]. Acoustic features associated with prosody include fundamental frequency, amplitude, duration, and segment quality [13], which are affected by the diverse impairments of speech musculature caused by dysarthria. This is important for communication because deficiencies in prosodic features interfere with the intelligibility of vowels, especially when tones are involved.

1.2. Phonatory Dysfunction

The speech impairments and abnormalities caused by dysarthria (see Figure 1) lead to phonatory dysfunction, which is one of the most frequently observed abnormalities perceived across all types of dysarthria [1, 2, 1416]. Phonatory dysfunction is a primary feature noted in clinical assessment of dysarthria [16]. Because a phoneme is generally regarded as an abstraction of a set of speech sounds (phones) which are perceived as equivalent to each other in a given language, phonatory dysfunction may be linked more generally with communication deficits [16].

Slow rate of speech and timing produce prolonged phonemes, and this can make a single-syllable word to be perceived (or recognized) as a two-syllable word (daydial). Also, words with long voiceless stops can be interpreted as two words because of the long silent occlusion phase in the middle of the target word (beforebe for) [17, 18].

In automatic speech recognition (ASR) and perception tests, phonatory dysfunction leads to an increase of deletion, insertion, and substitution of phonemes [7, 17, 19, 20]. These characteristics decrease the speaker’s intelligibility and restrict the speaker’s phonetic repertoire, causing that some sounds or phonemes cannot be uttered or articulated correctly.

1.3. Speech Recognition Technology

People with dysarthria also have muscular atrophy, which restricts their ability to use switches or keyboards for communication or control of assistive devices (e.g., an electric wheelchair). In this case, speech recognition technology is an attractive alternative for communication and control despite the difficulty of achieving robust recognition performance because of phonatory dysfunction.

Many assistive applications based on speech recognition have been developed and implemented. In [17, 2123] the use of commercial ASR systems for dictation (e.g., Dragon Naturally Speaking) was explored to improve speech communication. These studies showed varying levels of recognition performance in the range of 50% to 95% for users with different levels of dysarthria and small vocabularies (<50 words).

In contrast, in [19, 2428] specific ASR systems were developed for dysarthric speakers. In [24] an ASR system for dysarthric speakers was built with Artificial Neural Networks (ANNs). In comparison with a commercial ASR, the ANN-based system achieved higher recognition performance. In [25, 26] ASR systems built with Hidden Markov Models (HMMs) [29] achieved significant performances for Dutch and Japanese dysarthric speakers, respectively. In [27] a HMM-based ASR system was able to achieve recognition accuracies over 80% for British speakers with severe dysarthria and a restricted vocabulary (7–10 words) to control electronic devices (e.g., radio, TV). In [30], a hybrid approach that integrated HMMs and ANNs was presented to improve recognition of disordered speech. In [19], a HMM-based ASR system for dysarthric speakers was built to control a web browser with 47 pre-selected words achieving accuracies within the range of 34.3% to 83.3%. In [31], a HMM-based ASR was integrated with speech synthesis to improve intelligibility of dysarthric speech. Recognition accuracies of 65% to 80% were achieved with improvements on intelligibility as measured by the Mean Opinion Score (MOS). Finally, in [28, 32], the development of a HMM-based voice-input voice-output (VIVOCA) communication aid device for people with disordered speech was presented. Such device was intended to recognize and interpret a dysarthric person’s speech and generate a more intelligible (and clear) version of the same speech. This system achieved a mean accuracy of 67% in real usage situations with small vocabularies (14–47 words).

In general, ASR technologies are focused to identify (recognize) more accurately the sentences spoken by the dysarthric speaker independently of the severity of the dysarthria. This is very important for the development of applications (as those described above) which have the objective of improving communication and interaction with other people or other assistive systems.

1.4. Justification and Proposal

For people with dysarthric speech, the development of assistive technology based on ASR is dependent on the achievement of robust recognition performance. This is not an easy task because of the wide range of abnormalities associated to dysarthric speech. In addition, although high recognition performance is achieved (e.g., >80%), this is obtained after several speaker-specific training session. The associated costs, in money and time, cannot be afforded for most of the people with this disorder. Overcoming these difficulties is worthy as human-computer interaction using ASR is more suitable when the person with dysarthria also has muscular atrophy, which restricts the ability to use switches or keyboards. In such case, ASR interaction has been reported to be more efficient and faster than using mechanisms [23, 3335].

It is proposed that an ASR system that can learn the speech patterns with the less amount of training data is suitable for dysarthric speakers. A previous approach considered the response of the ASR system to estimate the phoneme confusion patterns of the speaker’s speech. Then, this information was integrated into the ASR process to correct those confusion errors (deletion, substitution, and/or insertion of phonemes) and provide a more accurate response [18, 26, 3638]. This approach performed better than other approaches that made use of speaker adaptation techniques (as those used by commercial ASR systems) because, as commented in [39], these are insufficient to deal with the abnormalities present in dysarthric speech. However, the performance of the confusion-matrix approach depends on the output of the ASR system and thus of its performance.

HMMs were considered for this work because these are the most frequent techniques used for recognition of normal and disordered speech. This is due to the efficiency of the HMMs to model the variation in the statistical properties of speech, both in the time and the frequency domains [40].

In this work, an approach based on finding suitable phoneme-specific HMM parameters for the ASR system is presented. It is argued that this approach can improve the acoustic modelling of phonemes affected by dysarthria and improve the performance of ASR, and thus, of other post-processing techniques as the phoneme confusion-matrix approach.

Based on the characteristics of dysarthric speech which were presented in this section the following parameters were considered for optimization of the ASR process.(i)Topology. This was considered an important parameter because in the works reviewed in Section 1.3 it was noted that for all phonemes or words in the ASR’s vocabulary the same HMM topology was used. As discussed in Sections 1.1 and 1.2, there is a significant variability in the phonetic dysfunction of vowels and consonants. Hence, specific topologies should be used for the acoustic modelling of each phoneme in the dysarthric speaker’s phonetic repertoire because not all phonemes are affected in the same way by dysarthria. For this work, the architecture (Bakis, Ergodic) and the number of states of the HMM were considered within the topology parameter. This is an extension on the work presented in [42], where topology optimization consisted in finding the optimal number of states for word HMMs considering just a Bakis left-to-right topology.(ii)Number of Gaussian Mixture Components. An important element of each state in a HMM are the emission probabilities. These represent the probabilities of an observation vector (e.g., the speech signal) being generated from a particular HMM state [43]. These probabilities are modelled by probability density functions (PDFs) which are represented as a weighted sum of Gaussians PDFs, each with different mean and variance [43, 44]. This is known as a Gaussian mixture, and the number of Gaussian PDFs affects the response of a HMM-based ASR system [4, 43, 45].

The selection of the suitable topologies and number of Gaussian mixture components for each phoneme in the dysarthric speaker’s language was performed with a Genetic Algorithm (GA) which is an important tool used in the field of optimization [46]. The performance of the ASR with the resulting GA-HMMs were compared with the approach of developing a speaker-dependent (SD) system, where training of HMMs is performed with the speech data of the target speaker [17, 27, 32, 35, 39]. The proposed approach achieved statistically significant gains on ASR accuracy when tested with the SD approach on a well-known database of dysarthric speech (Nemours [41]).

The details of the proposed phoneme-specific approach are presented in this paper as follows. In Section 2, the details about the selection of the HMM topologies and number of Gaussian mixture components for each phoneme are presented. Then, in Section 3, the information regarding the structure and elements of the GA used for the optimization of the phoneme’s HMM parameters is presented. In Section 4, the results obtained with the proposed approach and the comparison with the SD approach are presented. Finally, in Section 5 the results obtained are discussed and future work is presented.

2. HMM Parameters for Optimization

2.1. HMM Topology

An important element of the HMMs is the topology or structure. In Figures 2 and 3 the topologies of HMMs for recognition of phonemes are shown. These topologies are known as Bakis, and the most frequently used is the three-state left-to-right structure [29, 31, 41, 43, 45] of Figure 2. Nowadays, commercial ASR system are based on phoneme HMMs with Bakis structure.

Another topology is known as Ergodic which is shown in Figure 4. In comparison with the Bakis topology, in the Ergodic topology every state () can be reached from every other state in a finite number of steps [29]. For recognition of dysarthric speech, Ergodic is commonly used when the ASR system is based on whole-word recognition [7, 27, 47].

In the works reviewed in Sections 1.3 and 1.4, the structure of the HMMs is fixed for the modelling of all phonemes or words in the vocabulary of the speech application. In this work it is argued that for dysarthric speech the topology must be specific for each phoneme because of the effect of phonatory dysfunction and the wide range of abnormalities caused by the affected speech articulator. Hence, slow pronunciation that affects speaking rate and timing may be more evident for certain phonemes than for others. In this case, an Ergodic topology could be more suitable to model speech with inconsistencies in speaking rate instead of a Bakis topology.

On the other hand, a Bakis structure with a large number of states has been reported to be suitable for modelling of long and poorly differentiated phonetic units [41]. Because of this, the number of states is an important element to be considered in the topology of HMMs for recognition of dysarthric speech. The number varies from the standard three states [35, 45] to eight [41] and 11 states [32]. Thus, for the optimization of the topology the following levels were considered for each parameter: (i)type: Bakis of Figure 2 (Bakis-1), Bakis of Figure 3 (Bakis-2), and Ergodic of Figure 4 (three levels); (ii)number of states for each type: 3-to-11 (nine levels).

2.2. Gaussian Mixture Components

Another parameter considered for optimization is the number of Gaussian mixture components used for each HMM state. This parameter is important for the modelling of the emission probabilities, which represent the probabilities of an observation vector (e.g., the speech signal) being generated from a particular HMM state [43]. These probabilities are modelled as a weighted sum (mixture) of Gaussians PDFs, each with different mean and variance [43, 44].

The number of Gaussian PDFs (mixture components) affects the performance of a HMM-based ASR system [4, 43, 45]. High ASR performance is obtained with a number of Gaussian mixture components within the range of eight [38] to 16 [45]. However, usually less than ten Gaussian mixture components are used [35]. Thus, for the optimization of the number of mixture components 16 levels were considered based on these ranges 1-to-16.

In the following section, the details of the optimization method used for these parameters is presented.

3. Optimization Method: Genetic Algorithm

The selection of HMM topologies and number of mixture components is performed by a micro-Genetic Algorithm (micro-GA), which is a computational method based on Darwin’s rules of natural selection. A GA is a search heuristic that mimics the process of natural evolution and generates useful solutions for optimization problems [46].

In general, GAs have been used in the field of ASR research for the optimization of HMMs as presented in [42, 48, 49]. In [42, 49], a GA was used to optimize the observation probabilities and transition states for HMM-based ASR systems. In contrast with the proposed approach, in [42], the topology optimization consisted in finding the optimal number of states for word HMMs considering just the Bakis-1 left-to-right topology. The proposed approach extends on the work presented in [42] by eliminating the left-to-right restriction and considering other topologies as the Bakis-2 and Ergodic with more states. In Figure 5, the general structure of the micro-GA used for this work is presented.

The GA starts with an “Initial Population” of candidate solutions or “Individuals”. Each of these solutions is evaluated to assess its “Fitness” which is related to the problem to be solved. In this case, the problem consists in finding the assignation of topologies and number of mixture components that would increase ASR performance, and each individual represents a set of assignations. Then, fitness is evaluated as the recognition accuracy obtained with the assignations given by an individual.

These individuals are selected for “Reproduction” based on Darwin’s rule of “survival of the fittest” (e.g., the individuals with better fitness). It is expected that, as happens in nature, the individuals with better characteristics survive, reproduce, and produce “Offsprings” which inherit the characteristics of their “Parents” which are refined after some generations. Then, for this case, the individuals of the initial population become the parents for new solutions (offsprings) which are constructed by reproduction operators. It is expected that good solutions (assignation of topologies and number of mixture components) will produce better solutions. The selection of parents is performed based on their fitness.

After reproduction, the offsprings are evaluated to assess their fitness. If the offsprings are better than other individuals in the population then these will be replaced by them. The process iterates until no change in the overall fitness of the entire population is achieved (or after a given number of iterations).

In comparison with a conventional GA, a micro-GA can work with a very small initial population (typically four or five individuals [50]) which can be randomly generated. This algorithm can converge (e.g., to find an optimal solution) quickly within a few iterations and provide estimates as good as a conventional GA, where populations can be up to 1000 individuals. In the following sections the details of the micro-GA are presented.

3.1. Initial Population

The micro-GA starts with 10 individuals where (i)the 1st individual consists in the assignation of the Bakis topology of Figure 2 with three-states (Bakis-1) for all phonemes; (ii)the 2nd individual consists in the assignation of the Bakis topology of Figure 3 with four-states (Bakis-2) for all phonemes; (iii)the 3rd individual consists in the assignation of the Ergodic topology of Figure 4 with three-states for all phonemes; (iv)the 4th-to-10th individuals are randomly generated assignations of the Bakis and Ergodic topologies mentioned above with a number of states within the nine levels specified in Section 2.1.

In order to perform the reproduction of these individuals, the assignations are coded into “Chromosomes”, which are presented in Figure 6. Each solution is represented by a vector with 81 “Genes” or values where(i)from gene the numbers represent the topology to be assigned to the th phoneme (in this case, ): Bakis-1 , Bakis-2 , Ergodic ; (ii)from gene the numbers represent the number of HMM states considered for the topology assigned to the th phoneme (3–11); (iii)gene 81 represents the number of Gaussian mixture components used for acoustic modelling with the HMMs (1–16).

3.2. Fitness Evaluation

For each individual in the population, a set of HMMs is built with the assigned topologies. The parameters of these HMMs as emission and transition probabilities are estimated with the Baum-Welch and Viterbi algorithms [43]. This process, called supervised training, is performed with a set of “training” speech.

Then, to measure the “Fitness” of each individual, the % Word Recognition Accuracy (WAcc) is computed on a set of “testing” speech. This measure is computed as where is the number of elements (words or phonemes) in the correct transcription of the spoken speech, and , and are the number of elements deleted, substituted, and inserted in the output generated by the HMM-based ASR system when compared to the correct transcription.

3.3. Selection of Parents

For this process, the Roulette Wheel selection [51] was performed as follows. (1) For each individual in the population its fitness is computed as specified in Section 3.2. (2) Compute the selection probability for each individual as If there is a situation where negative values are obtained, the most negative is taken as reference. Then, the absolute value of the reference is added to all values in the population. In this way, the most negative value gets a fitness of 0, and the individuals with less negative get new positive (but small) values. The individuals with positive get their fitness increased accordingly to the absolute of the reference value. This adjustment does not change the concept of fitness as a value of 0 represents an individual with very poor abilities to solve the problem.(3) Compute the cumulative probability for each individual as . (4) Generate a uniform random number . (5) If , then select the first individual (), otherwise, select such that . (6) Repeat Steps 4 and 5 times until all individuals are selected.

This procedure gives as output pairs or couples of parents which then can produce offsprings by means of the reproduction operators known as crossover and mutation.

3.4. Reproduction of Parents

The reproduction operators enable the creation of offsprings (new solutions) from an initial set of individuals (parents). This process is equivalent to exploring points within the solution space of a problem: parents are initial solutions for the problem, and creating offsprings is equivalent to finding other solutions for the same problem.

The first reproduction operator is known as “Crossover” and consists in the interchange of genes between the parent’s chromosomes [46]. Crossover is explorative as it makes a jump to a region somewhere “in between” two (parents) regions [52]. Because of this, the crossover operator diversifies the population [53].

There are many crossover schemes, and the use of a particular scheme depends of the kind of choromosome codification and type of problem to be solved. In this case, the chromosome represents assignations that represent topologies and number of Gaussian mixture components, and the values of each gene are positive integer numbers. For this codification, the linear crossover was used [5456]. Considering two parent choromosomes, and , where and are the th genes and is the length of the chromosome, the genes for the offsprings and are obtained as where is a weight value which in this case was associated to the crossover probability. In this way, each th gene of an offspring is a new value created from the arithmetic combination of genes at the same th position of the parent chromosomes [56]. Note that, from the selection method presented in Section 3.3, parents form couples, and each couple produces two offsprings; thus, offsprings are created with the crossover operator.

The second reproduction operator is known as “Mutation” and consists in changing (randomly or based on a probability) a number of genes across all individuals. Thus, the mutation operator can create new individuals by making changes in a single individual. Mutation is exploitative as it creates random small deviations, thereby staying near (in the region of) the parent [52]. While the crossover operator diversifies the population, the mutation operator exploits the new result [53].

The changes performed by the mutation operator may consist in just selecting a gene and assigning it a different value within the associated allowable range, or selecting two genes and interchange their values. For this case, a number of parents are randomly selected from the initial population. Then, for each selected parent, the values of randomly selected genes are changed with values within the associated ranges. The and numbers are associated with the mutation probability as follows:

This leads to changed parents that become the offsprings generated by mutation. In this work, the probabilities of both operators are related based on the following equivalence:

The probability of mutation was set as the reference for the crossover probability. The probability of mutation was considered to be increased depending on the number of iterations or generations of the GA. This increment was considered according to the expression which is plotted in Figure 7. For this expression, represents the percentage of the number of iterations of the micro-GA.

From Figure 7, four values for are considered: 0.10, 0.30, 0.40, and 0.50. If is defined as the number of iterations of the micro-GA, these values are considered when the algorithm reaches = 10%, 30%, 50%, and 70% of , respectively. This form of estimation for was considered in order to dynamically change the intensity of the explorative and exploitative searching process performed with the reproduction operators.

Initially, at iteration = 0, , thus and just crossover is performed for all individuals in the initial population. When the GA reaches, 10% of the total number of iterations (), and , thus crossover is performed on individuals and mutation is performed on individuals. This value for is kept until the GA reaches 30% of , where and , leading to crossover being performed on individuals and mutation being performed on individuals. Then, this crossover and mutation rate are kept until the GA reaches 50% of , where and . This continues until the GA reaches 70% of , where .

In this way, at the beginning of the GA explorative searching is mainly performed, thus intensifying diversification. As the GA continues, the exploitative search is increased until both are performed with the same intensity. Finally, the levels considered for and are consistent with other GA implementations [53, 55].

3.5. Stop Condition

As commented in [52], there are many stop criteria for a GA, like considering a maximum number of generations or iterations, a maximum number of functional evaluations, or convergence is achieved. For the micro-GA, a fixed number of iterations was considered, in this case . This is consistent with the stop condition of the GA presented in [42] (stop after 30 iterations). In experiments, however, it was observed that changes in convergence were minimal after 20 iterations.

4. Experiments on Dysarthric Speech

4.1. Speech Data

For the experiments the Nemours database of dysarthric speech was used [41]. This database has been widely used in ASR research as presented in [31, 36, 45, 57, 58]. The Nemours database consists of speech data from ten American-English speakers with dysarthrias resulting from either cerebral palsy or head trauma with associated quadriplegia [41]. The main speech data consists of a collection of 74 short sentences spoken by each speaker (740 sentences in total). These sentences are nonsense phrases that have a simple syntax of the form “the is the ”, where and are monosyllabic nouns (74 in total) and is a bisyllabic verb (37 in total) in present participle form. Specific sentences were generated by randomly selecting and () without replacement from the set of 74 nouns, and selecting without replacement from the set of 37 verbs. This process produced the first 37 sentences, and the other 37 sentences were generated by swapping the and words in the first set. Because of this, in the complete set of 74 sentences there are two pronunciations of each noun and verb. The vocabulary in this set consists of 111 different words.

With this speech data, an initial assessment and recognition test were performed with human listeners. Because this paper is focused on the recognition task, the recognition scores produced by the human listeners for each dysarthric speaker are presented in Table 1. This information is important to identify the speakers with severe, moderate and mild levels of dysarthria based on their recognition scores. More information about this test and the intelligibility assessment for these speakers can be found in [41, 59].

Talker Average Talker Average

BB 89.7 LL 84.4
BK 58.2 MH 92.1
BV 57.5 RK 68.6
FB 92.9 RL 73.3
JF 78.5 SC 51.5

Based on the data presented in Table 1 and the identification presented in [45], the speakers were classified as presented in Table 2. In [45], four speakers were considered as moderate and three as mild. In this work, a speaker from the moderate group was taken to the mild group according to the data presented in Table 1 [41]. This classification is important for the training scheme of the SD systems presented in Section 4.2.

Speaker Level of dysarthria Speaker Level of dysarthria

FB Mild RL Moderate
MH Mild RK Moderate
BB Mild BK Severe
LL Mild BV Severe
JF Moderate SC Severe

In addition, two readings of narratives identified as “My Grandfather” and “The Rainbow” are included per speaker. While the sets of 74 sentences are phonetically and orthographically labelled, the narratives are not labelled at any level. Thus, extra labelling was performed to consider this speech material. The narratives were separated into sentences, leading to 18 sentences for the “My Grandfather” narrative, and 14 sentences for the “The Rainbow” narrative. The vocabulary in these narratives consisted of 158 different words. In Table 3 the selection of sentences used for training, fitness evaluation, and testing of the GA-HMMs is presented. This selection was defined to include all phonemes present in the speech database in the training, fitness evaluation, and testing sets.



Fitness evaluation5105131115193337


All speech data was coded into MFCC format where the front-end used 12 MFCCs plus energy, delta and acceleration coefficients. Also a frame period of 10 msec with a Hamming window of 25 msec and 26 filter-bank channels were used [43].

4.2. ASR Systems

The implementation tool for the HMMs and the recognition tasks was performed with HTK [43]. Because in the Nemours database 40 phonemes were identified, 40 monophone acoustic HMMs were constructed for each type of ASR system. In this work, two ASR systems were considered.(1)Speaker-Dependent (SD) ASR. The HMMs have the same parameters (topology, number of states, number of Gaussian mixture components per state) for all phonemes, and these are trained (built) with speech data of the target (test) speaker. This is the common approach for the development of ASR for dysarthric speech as performed in [17, 27, 32, 35, 39]. Thus, this ASR provides the baseline (or reference) performance for comparison purposes. About the parameters for this SD ASR, in Section 3.1 was determined that the first three individuals in the initial population for the micro-GA were ASR systems built with the topologies presented in Figures 2, 3, and 4, covering standard Bakis and Ergodic topologies. Hence, three baseline SD ASR systems were built as reference systems: SD Bakis-1, SD Bakis-2, and SD Ergodic.Additionally, two schemes for building the SD ASR were considered: using the training data from all speakers, including the target (test) speaker and using the training data from just the target (test) speaker. Note that, if under the scheme no speech data from the target speaker were used, then the system would be completely speaker-independent (SI), and thus, an adaptation technique would be required. In Table 4 the percentage of recognition accuracy obtained by the SD ASR systems under these two training schemes on the testing sets across all speakers is presented.As presented, on average, speakers with mild to moderate dysarthria achieve higher recognition performance when training of the SD also includes speech data from other speakers. However, speakers with more severe dysarthria achieve better performance when the SD is trained only with speech data of the target speaker. This situation was also observed in [17, 45]. Because of these results, dysarthric-specific schemes were considered for training of the SD ASR systems: (i) mild scheme: training speech data from speakers FB, MH, BB, LL was used to train the SD for the same speakers; (ii) moderate scheme: training speech data from speakers JF, RL, RK was used to train the SD for the same speakers; (iii) severe scheme: training speech data from speakers BK, BV, SC was used to train the SD for the same speakers. In Table 5, the recognition results obtained with the dysarthric-specific schemes are presented. In Figure 8, all schemes are presented for comparison. With the dysarthric-specific training schemes it was possible to achieve similar performance (and in some cases, higher performance) when compared with the schemes presented in Table 4. This is important because an ASR can be built for specific categories or levels of dysarthria. This may be the reason why, as observed in [17, 45], speakers with mild dysarthria get more benefits from using ASR built with normal speech (adapted speaker-independent ASR system) than by using SD ASR systems (in this case, mild dysarthric speakers are closer to normal speech than moderate or severe dysarthric speakers). Thus, for the optimization of the HMMs, the dysarthric-specific scheme was used for the creation of the baseline SD ASR systems.(2)GA-Optimized (GA-op) SD ASR. The HMMs of the baseline SD ASR are optimized with the micro-GA presented in Section 3 to make each HMM specific to the characteristics of the phoneme. Optimization involves the identification of the most suitable topology and number of Gaussian mixture components to improve the performance of the ASR system. For the micro-GA, training-independent speech data was used for fitness evaluation (see Table 3).

SpeakerTraining scheme 1
SD Bakis-1SD Bakis-2SD ErgodicAverage scheme 1


SpeakerTraining scheme 2
SD Bakis-1SD Bakis-2SD ErgodicAverage scheme 2


SpeakerSD Bakis-1SD Bakis-2SD ErgodicAverage

Mild scheme

Moderate scheme

Severe scheme

4.3. Convergence of the Micro-GA

In Figure 9, the average convergence plot of the GA-op SD HMMs across all iterations of the micro-GA on the fitness evaluation set is presented. This illustrates that the assignment of different topologies can lead to improvement on ASR performance. In Table 6, the assignations obtained with the micro-GA for each speaker’s set of phonemes are presented.



No. Bakis-13740404040133736340
No. Bakis-2200001432300
No. Ergodic10000130270

Average No. States5.333.154.454.334.186.883.306.407.955.48
Standard Deviation2.140.661.040.620.712.551.181.552.771.04

No. Gaussian Mixtures666671266167

4.4. Performance of the GA-op ASR System

In Table 7 and Figure 10 the comparison of performances of the baseline SD ASR systems with the GA-op SD ASR is presented. The GA-op was compared with the baseline SD ASR system with the higher performance from Table 5. As presented, the performance of the GA-op SD ASR system is higher than the baseline’s for each speaker independently of the dysarthric level.

Level of dysarthriaSpeakerBaseline SD ASRGA-op ASR
Type% WAcc% WAcc

MildFBSD Bakis-1 (3, 6)81.2284.72
MHSD Bakis-1 (3, 6)80.3583.41
BBSD Bakis-1 (3, 6)77.5279.82
LLSD Bakis-1 (3, 6)74.1177.23

ModerateJFSD Bakis-1 (3, 6)64.0066.67
RLSD Bakis-1 (3, 6)51.1160.00
RKSD Ergodic (3, 6)39.9147.37

SevereBKSD Bakis-1 (3, 6)36.7346.46
BVSD Bakis-1 (3, 6)56.8261.82
SCSD Bakis-1 (3, 6)47.2754.55

On average, an increase of 5.3% was achieved across all dysarthric speakers (66.20%–60.90%). These results were statistically significant with a value <0.10 using the matched-pairs test described in [60]. It is acknowledged that, in practical terms, ASR should be higher as is for normal speech (80%–96% for small vocabularies) [61]; however, human recognition for dysarthric speech has been reported to be accurate between 7% and 61% of the time [45, 62]. The performance achieved with the proposed approach is higher than human recognition based on this information.

5. Discussion and Future Work

In this work an approach consisting of modelling each phoneme with a specific topology for dysarthric speech was presented. Initially the approach of developing SD ASR systems was considered. Results presented in Figure 8 showed that by adding speech samples from other speakers with similar level of dysarthria to the training of the SD ASR system, higher performance could be achieved for some speakers. As presented in Table 7 and Figure 10, when optimizing the topologies of the HMMs of this SD ASR system more gains in recognition performance were obtained. In practice, this approach can be applied if the speaker already has a SD ASR system, or for the designing of dysarthric-specific SD ASR systems. In such cases, an automatic trainer and builder of ASR systems must be developed. In the field of human-computer interaction (HCI), this approach can be used to further improve the efficiency of assistive interfaces.

From Table 6 it is observed that the Bakis-1 topology is the most suitable for all speakers with mild dysarthria, and just for some with moderate and severe dysarthria. In Particular for speaker RL the topologies Bakis-1, Bakis-2, and Ergodic were equally assigned to different phonemes with an average of seven states and 12 Gaussian mixture components per state. This speaker was classified with moderate dysarthria based on the recognition results presented in Table 2. However the assignations estimated by the micro-GA indicate that this speaker may present a wide range of variations in the pronunciation of phonemes (and, thus, more specific HMMs must be considered). Note than an improvement of almost 10% was obtained for speaker RL after the optimization of the HMMs of the baseline SD ASR system.

For speakers BK and BV, which were classified with severe dysarthria, the most suitable topologies were Bakis-1 and Bakis-2 (with some Ergodic), respectively. However note that the average number of states is significantly higher (seven and eight resp.) than for the mild and moderate speakers (excluding speaker RL). For speaker BV, the number of Gaussian mixture components is the highest with 16.

In general, speakers RL, BK, and BV required more states for the acoustic modelling of their phonemes. Speakers FB, MH, BB, LL (mild dysarthria), and JF (moderate dysarthria) required less states, with an average of four-to-six states. Speaker RK, which was classified with moderate (but close to severe) dysarthria, required less states with an average of four. Speaker SC, which is the one with the lowest recognition performance in Table 2 and was classified with severe dysarthria, required an average of six states.

Although for speakers FB, MH, BB, LL, JF, RL, BK, and BV clear assignations were observed based on their level of dysarthria (e.g., mild to severe dysarthria leads to increase the number of HMM states for acoustic modelling), for speakers RK and SC these were not observed (e.g., assignations for the moderate-to-severe and severe RK and SC speakers are more associated to a mild dysarthric speaker). However, the improvements achieved with the proposed approach was consistent across all speakers and levels of dysarthria.

Overall, the standard three-state left-to-right Bakis topology used for acoustic modelling of phonemes requires more states for the modelling of dysarthric speech. And in some cases, more than one type of topology is required. Also, this is dependent of the acoustic characteristics of each phoneme in the speaker’s repertoire.

Future work is focused on extending the study of the situations observed in this work:(i)to explore on the use of dynamic topologies, where besides changing the number of states, the transitions between them can also be changed; (ii)to improve the convergence of the micro-GA with alternative crossover and mutation operators; (iii)to test the approach on a different and larger database of dysarthric speech (e.g., the TORGO database); (iv)to incorporate the post-processing confusion-matrix approach presented in [18, 26, 38] for further improvement (for this, more dysarthric speech data would be required); (v)to explore on the use of the HMM assignations for the assessment of dysarthric speech.


  1. F. L. Darley, A. E. Aronson, and J. R. Brown, “Differential diagnostic patterns of dysarthria,” Journal of Speech and Hearing Research, vol. 12, no. 2, pp. 246–269, 1969. View at: Google Scholar
  2. F. L. Darley, A. E. Aronson, and J. R. Brown, “Clusters of deviant speech dimensions in the dysarthrias,” Journal of Speech and Hearing Research, vol. 12, no. 3, pp. 462–496, 1969. View at: Google Scholar
  3. A. B. Kain, J.-P. Hosom, X. Niu, J. P. H. van Santen, M. Fried-Oken, and J. Staehely, “Improving the intelligibility of dysarthric speech,” Speech Communication, vol. 49, no. 9, pp. 743–759, 2007. View at: Publisher Site | Google Scholar
  4. H. V. Sharma, Acoustic model adaptation for recognition of dysarthric speech [Ph.D. dissertation], University of Illinois, Urbana, Ill, USA, 2012.
  5. P. C. Doyle, H. A. Leeper, A. L. Kotler et al., “Dysarthric speech: a comparison of computerized speech recognition and listener intelligibility,” Journal of Rehabilitation Research and Development, vol. 34, no. 3, pp. 309–316, 1997. View at: Google Scholar
  6. A. Kain, X. Niu, J. P. Hosom, Q. Miao, and J. P. H. van Santen, “Formant re-synthesis of dysarthric speech,” in Proceedings of the ISCA Speech Synthesis Workshop, 2004. View at: Google Scholar
  7. P. D. Polur and G. E. Miller, “Effect of high-frequency spectral components in computer recognition of dysarthric speech based on a Mel-cepstral stochastic model,” Journal of Rehabilitation Research and Development, vol. 42, no. 3, pp. 363–372, 2005. View at: Publisher Site | Google Scholar
  8. W. J. Hardcastle, R. A. Morgan Barry, and C. J. Clark, “Articulatory and voicing characteristics of adult dysarthric and verbal dyspraxic speakers: an instrumental study,” British Journal of Disorders of Communication, vol. 20, no. 3, pp. 249–270, 1985. View at: Google Scholar
  9. G. Weismer, “Articulatory characteristics of Parkinsonian dysarthria: segmental and phrase-level timing, spirantization, and glottal-supraglottal coordination,” in The Dysarthrias: Physiology, Acoustics, Perception, Management, M. R. McNeil, J. C. Rosenbek, and A. E. Aronson, Eds., pp. 101–130, Timonium, Maryland, Md, USA, College Hill Press, San Diego, Calif, USA. View at: Google Scholar
  10. R. D. Kent and J. C. Rosenbek, “Acoustic patterns of apraxia of speech,” Journal of Speech and Hearing Research, vol. 26, no. 2, pp. 231–249, 1983. View at: Google Scholar
  11. H. Ackermann and I. Hertrich, “Dysarthria in Friedreich's ataxia: timing of speech segments,” Clinical Linguistics and Phonetics, vol. 7, no. 1, pp. 75–91, 1993. View at: Google Scholar
  12. H. Kim, H. Hasegawa-Johnson, and A. Perlman, “Acoustic cues to lexical stress in spastic dysarthria,” in Proceedings of the Speech Prosody, vol. 100891, pp. 1–4, 2010. View at: Google Scholar
  13. R. Patel, “Prosodic control in severe dysarthria: preserved ability to mark the question-statement contrast,” Journal of Speech, Language, and Hearing Research, vol. 45, no. 5, pp. 858–870, 2002. View at: Google Scholar
  14. R. D. Kent, J. F. Kent, J. R. Duffy, and G. Weismer, “The dysarthrias: speech-voice profiles, related dysfunctions, and neuropathology,” Journal of Medical Speech-Language Pathology, vol. 6, no. 4, pp. 165–211, 1998. View at: Google Scholar
  15. W. Ziegler and P. Hoole, “Voice quality measurement,” in Neurologic Disease, R. D. Kent and M. J. Ball, Eds., pp. 397–410, Singular, 2000. View at: Google Scholar
  16. R. D. Kent, H. K. Vorperian, J. F. Kent, and J. R. Duffy, “Voice dysfunction in dysarthria: application of the multi-dimensional voice program,” Journal of Communication Disorders, vol. 36, no. 4, pp. 281–306, 2003. View at: Publisher Site | Google Scholar
  17. P. Raghavendra, E. Rosengren, and S. Hunnicutt, “An investigation of different degrees of dysarthric speech as input to speaker-adaptive and speaker-dependent recognition systems,” Augmentative and Alternative Communication, vol. 17, no. 4, pp. 265–275, 2001. View at: Google Scholar
  18. S. O. Caballero Morales and S. J. Cox, “Modelling errors in automatic speech recognition for dysarthric speakers,” EURASIP Journal on Advances in Signal Processing, vol. 2009, Article ID 308340, 14 pages, 2009. View at: Publisher Site | Google Scholar
  19. F. Hamidi, M. Baljko, N. Livingston et al., “A customizable speech interface for people with dysatric speech,” in Proceedings of the 12th International Conference on Computers Helping People with Special Needs (ICCHP '10), K. Miesenberger, J. Klaus, W. Zagler, and A. Karshmer, Eds., vol. 6179 of Lecture Notes in Computer Sciences (LNCS), pp. 605–612, Springer, Berlin, Germany, 2010. View at: Google Scholar
  20. K. Rosen and S. Yampolsky, “Automatic speech recognition and a review of its functioning with dysarthric speech,” Augmentative and Alternative Communication, vol. 16, no. 1, pp. 48–60, 2000. View at: Google Scholar
  21. L. Ferrier, H. Shane, H. Ballard, T. Carpenter, and A. Benoit, “Dysarthric speaker's intelligibility and speech characteristics in relation to computer speech recognition,” Augmentative and Alternative Communication, vol. 11, no. 3, pp. 165–175, 1995. View at: Publisher Site | Google Scholar
  22. N. J. Manasse, K. Hux, and J. L. Rankin-Erickson, “Speech recognition training for enhancing written language generation by a traumatic brain injury survivor,” Brain Injury, vol. 14, no. 11, pp. 1015–1034, 2000. View at: Google Scholar
  23. N. Manasse, K. Hux, J. Rankin-Erickson, and E. Lauritzen, “Accuracy of three speech recognition systems: case study of dysarthric speech,” Augmentative and Alternative Communication, vol. 16, no. 3, pp. 186–196, 2000. View at: Google Scholar
  24. G. Jayaram and K. Abdelhamied, “Experiments in dysarthric speech recognition using artificial neural networks,” Journal of Rehabilitation Research and Development, vol. 32, no. 2, pp. 162–169, 1995. View at: Google Scholar
  25. H. Strik, E. Sanders, M. Ruiter, and L. Beijer, “Automatic recognition of dutch dysarthric speech: a pilot study,” in Proceedings of the 7th International Conference on Spoken Language Processing (ICSLP '02), pp. 661–664, 2002. View at: Google Scholar
  26. H. Matsumasa, T. Takiguchi, Y. Ariki, I. Li, and T. Nakabayashi, “Integration of metamodel and acoustic model for speech recognition,” in Proceedings of the 9th Annual Conference of the International Speech Communication Association (Interspeech '08), pp. 2234–2237, September 2008. View at: Google Scholar
  27. M. S. Hawley, P. Enderby, P. Green et al., “A speech-controlled environmental control system for people with severe dysarthria,” Medical Engineering & Physics, vol. 29, no. 5, pp. 586–593, 2007. View at: Publisher Site | Google Scholar
  28. M. S. Hawley, P. Enderby, P. Green, S. Cunningham, and R. Palmer, “Development of a voice-input voice-output communication aid (VIVOCA) for people with severe dysarthria,” in Proceedings of the 10th International Conference on Computers Helping People with Special Needs (ICCHP '06), K. Miesenberger, J. Klaus, W. L. Zagler, and A. I. Karshmer, Eds., vol. 4061 of Lecture Notes in Computer Science (LNCS), pp. 882–885, 2006. View at: Google Scholar
  29. L. R. Rabiner, “A tutorial on hidden Markov models and selected applications in speech recognition,” Proceedings of the IEEE, vol. 77, no. 2, pp. 257–286, 1989. View at: Publisher Site | Google Scholar
  30. P. D. Polur and G. E. Miller, “Investigation of an HMM/ANN hybrid structure in pattern recognition application using cepstral analysis of dysarthric (distorted) speech signals,” Medical Engineering & Physics, vol. 28, no. 8, pp. 741–748, 2006. View at: Publisher Site | Google Scholar
  31. M. S. Yakcoub, S.-A. Selouani, and D. O'Shaughnessy, “Speech assistive technology to improve the interaction of dysarthric speakers with machines,” in Proceedings of the 3rd International Symposium on Communications, Control, and Signal Processing (ISCCSP '08), pp. 1150–1154, March 2008. View at: Publisher Site | Google Scholar
  32. M. S. Hawley, P. Enderby, P. Green et al., “A voice-input voice-output communication aid for people with severe speech impairment,” IEEE Transactions on Neural Systems and Rehabilitation Engineering, vol. 21, no. 1, pp. 23–31, 2013. View at: Publisher Site | Google Scholar
  33. H.-P. Chang, “Speech input for dysarthric users,” in Proceedings of the Meeting of the Acoustical Society of America, 1993. View at: Google Scholar
  34. N. Thomas-Stonell, A.-L. Kotler, H. A. Leeper, and P. C. Doyle, “Computerized speech recognition: influence of intelligibility and perceptual consistency on recognition accuracy,” Augmentative and Alternative Communication, vol. 14, no. 1, pp. 51–56, 1998. View at: Google Scholar
  35. M. Hasegawa-Johnson, J. Gunderson, A. Perlman, and T. Huang, “HMM-based and SVM-based recognition of the speech of talkers with spastic dysarthria,” in Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '06), vol. 3, pp. 1060–1063, May 2006. View at: Google Scholar
  36. W. K. Seong, J. H. Park, and H. K. Kim, “Dysarthric speech recognition error correction using weighted finite state transducers based on context-dependent pronunciation variation,” in Proceedings of the 13th International Conference on Computers Helping People with Special Needs (ICCHP '12), K. Miesenberger, A. I. Karshmer, P. Penaz, and W. L. Zagler, Eds., vol. 7383 of Lecture Notes in Computer Science, pp. 475–482, Springer, 2012. View at: Google Scholar
  37. W. K. Seong, J. H. Park, and H. K. Kim, “Multiple pronunciation lexical modeling based on phoneme confusion matrix for dysarthric speech recognition,” Advanced Science and Technology Letters, vol. 14, pp. 57–60, 2012. View at: Google Scholar
  38. S. O. Caballero-Morales and F. Trujillo-Romero, “Dynamic estimation of phoneme confusion patterns with a genetic algorithm to improve the performance of metamodels for recognition of disordered speech,” in Advances in Computational Intelligence, I. Batyrshin and M. González-Mendoza, Eds., vol. 7630 of Lecture Notes in Artificial Intelligence, pp. 175–186, Springer, 2013. View at: Google Scholar
  39. P. Green, J. Carmichael, A. Hatzis, P. Enderby, M. S. Hawley, and M. Parker, “Automatic speech recognition with sparse training data for dysarthric speakers,” in Proceedings of the the 8th European Conference on Speech Communication and Technology (EUROSPEECH '03), pp. 1189–1192, 2003. View at: Google Scholar
  40. M. Frikha and A. B. Hamida, “A comparative survey of ANN and hybrid HMM/ANN architectures for robust speech recognition,” American Journal of Intelligent Systems, vol. 2, no. 1, pp. 1–8, 2012. View at: Google Scholar
  41. X. Menéndez-Pidal, J. B. Polikoff, S. M. Peters, J. E. Leonzio, and H. T. Bunnell, “The nemours database of dysarthric speech,” in Proceedings of the 4th International Conference on Spoken Language Processing (ICSLP '96), vol. 3, pp. 1962–1965, October 1996. View at: Google Scholar
  42. C. W. Chau, S. Kwong, K. F. Man, and K. S. Tang, “Optimisation of HMM topology and its model parameters by genetic algorithms,” Pattern Recognition, vol. 34, no. 2, pp. 509–522, 2001. View at: Publisher Site | Google Scholar
  43. S. Young and P. Woodland, The HTK Book, (for HTK Version 3.4), Cambridge University Engineering Department, 2006.
  44. D. Jurafsky and J. H. Martin, Speech and Language Processing, Pearson Prentice Hall, 2009.
  45. F. Rudzicz, “Comparing speaker-dependent and speaker-adaptive acoustic models for recognizing dysarthric speech,” in Proceedings of the 9th International ACM SIGACCESS Conference on Computers and Accessibility (ASSETS '07), pp. 255–256, October 2007. View at: Publisher Site | Google Scholar
  46. D. E. Goldberg, Genetic Algorithms in Search, Optimization and Machine Learning, Addison-Wesley, 1989.
  47. L. J. Ferrier, J. R. Deller, and D. Hsu, “On the use of hidden Markov modelling for recognition of dysarthric speech,” Computer Methods and Programs in Biomedicine, vol. 35, no. 2, pp. 125–139, 1991. View at: Publisher Site | Google Scholar
  48. Q. Y. Hong and S. Kwong, “A genetic classification method for speaker recognition,” Engineering Applications of Artificial Intelligence, vol. 18, no. 1, pp. 13–19, 2005. View at: Publisher Site | Google Scholar
  49. T. Takara, Y. Iha, and I. Nagayama, “Selection of the optimal structure of the continuous HMM using the genetic algorithm,” in Proceedings of the 5th International Conference on Spoken Language Processing (ICSLP '98), 1998. View at: Google Scholar
  50. G. A. Bakare, G. K. Venayagamoorthy, and U. O. Aliyu, “Reactive power and voltage control of the Nigerian grid system using micro-genetic algorithm,” in Proceedings of IEEE Power Engineering Society General Meeting, vol. 2, pp. 1916–1922, June 2005. View at: Google Scholar
  51. K. F. Leung, F. H. F. Leung, H. K. Lam, and S. H. Ling, “Application of a modified neural fuzzy network and an improved genetic algorithm to speech recognition,” Neural Computing and Applications, vol. 16, no. 4-5, pp. 419–431, 2007. View at: Publisher Site | Google Scholar
  52. B. Kumar and R. Dhiman, “Tuning of PID controller for liquid level tank system using intelligent techniques,” International Journal of Computer Science and Technology, vol. 2, no. 4, pp. 257–260, 2011. View at: Google Scholar
  53. R. Kumar, “An experimental analysis of explorative and exploited operators of genetic algorithm for operating system process scheduling problem,” International Journal of Engineering and Technology, vol. 2, no. 6, pp. 472–476, 2010. View at: Google Scholar
  54. T. Nomura, “Analysis on linear crossover for real number chromosomes in an infinite population size,” in Proceedings of IEEE International Conference on Evolutionary Computation (ICEC '97), pp. 111–114, April 1997. View at: Google Scholar
  55. J. Xiao, L. Zou, and C. Li, “Optimization of hidden Markov model by a genetic algorithm for web information extraction,” in Proceedings of the International Conference on Intelligent Systems and Knowledge Engineering (ISKE '07), 2007. View at: Google Scholar
  56. D. T. Vollmer, T. Soule, and M. Manic, “A distance measure comparison to improve crowding in multi-modal optimization problems,” in Proceedings of the 3rd International Symposium on Resilient Control Systems (ISRCS '10), pp. 31–36, August 2010. View at: Publisher Site | Google Scholar
  57. J. P. Hosom, A. B. Kain, T. Mishra, J. P. H. van Santen, M. Fried-Oken, and J. Staehely, “Intelligibility of modifications to dysarthric speech,” in Proceedings of IEEE International Conference on Accoustics, Speech, and Signal Processing (ICASSP '03), vol. 1, pp. 924–927, April 2003. View at: Publisher Site | Google Scholar
  58. W. K. Seong, J. H. Park, and H. K. Kim, “Performance improvement of dysarthric speech recognition using context-dependent pronunciation variation modeling based on Kullback-Leibler distance,” Advanced Science and Technology Letters, vol. 14, pp. 53–56, 2012. View at: Google Scholar
  59. H. T. Bunnel and J. B. Polikoff, “The nemours database of dysarthric speech: a perceptual analysis,” in Proceedings of the 14th International Congress of Phonetic Sciences, vol. 1, pp. 783–786, 1999. View at: Google Scholar
  60. L. Gillick and S. J. Cox, “Some statistical issues in the comparison of speech recognition algorithms,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '89), pp. 532–535, May 1989. View at: Google Scholar
  61. National Institute of Standards and Technology (NIST), The History of Automatic Speech Recognition Evaluations at NIST, 2009,
  62. J. M. Noyes and C. R. Frankish, “Speech recognition technology for individuals with disabilities,” Augmentative and Alternative Communication, vol. 8, no. 4, pp. 297–303, 1992. View at: Publisher Site | Google Scholar

Copyright © 2013 Santiago-Omar Caballero-Morales. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

1801 Views | 584 Downloads | 2 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder