International Journal of Vehicular Technology

International Journal of Vehicular Technology / 2013 / Article
Special Issue

Advances of Human Factors Research for Future Vehicular Technology

View this Special Issue

Review Article | Open Access

Volume 2013 |Article ID 924170 |

Victor Ei-Wen Lo, Paul A. Green, "Development and Evaluation of Automotive Speech Interfaces: Useful Information from the Human Factors and the Related Literature", International Journal of Vehicular Technology, vol. 2013, Article ID 924170, 13 pages, 2013.

Development and Evaluation of Automotive Speech Interfaces: Useful Information from the Human Factors and the Related Literature

Academic Editor: Motoyuki Akamatsu
Received11 Oct 2012
Revised10 Jan 2013
Accepted17 Jan 2013
Published07 Mar 2013


Drivers often use infotainment systems in motor vehicles, such as systems for navigation, music, and phones. However, operating visual-manual interfaces for these systems can distract drivers. Speech interfaces may be less distracting. To help designing easy-to-use speech interfaces, this paper identifies key speech interfaces (e.g., CHAT, Linguatronic, SYNC, Siri, and Google Voice), their features, and what was learned from evaluating them and other systems. Also included is information on key technical standards (e.g., ISO 9921, ITU P.800) and relevant design guidelines. This paper also describes relevant design and evaluation methods (e.g., Wizard of Oz) and how to make driving studies replicable (e.g., by referencing SAE J2944). Throughout the paper, there is discussion of linguistic terms (e.g., turn-taking) and principles (e.g., Grice’s Conversational Maxims) that provide a basis for describing user-device interactions and errors in evaluations.

1. Introduction

In recent years, automotive and consumer-product manufacturers have incorporated speech interfaces into their products. Published data on the number of vehicles sold with speech interfaces is not readily available, though the numbers appear to be substantial. Speech interfaces are of interest because visual-manual alternatives are distracting, causing drivers to look away from the road, and increasing crash risk. Stutts et al. [1] reported that adjusting and controlling entertainment systems and climate-control systems and using cell phones accounted for 19% of all crashes related to distraction. The fact that the use of entertainment systems is ranked the second among major causes of these crashes arises the argument that speech interfaces should be used for music selection. Tsimhoni et al. [2] reported that 82% less time was needed for drivers to enter an address using a speech interface as opposed to using a keyboard, indicating that a speech interface is preferred for that task. However, using a speech interface still requires cognitive demand, which can interfere with the primary driving task. For example, Lee et al. [3] showed that drivers’ reaction time increased by 180 ms when using a complex speech-controlled email system (three levels of menus with four-to-seven options for each menu) in comparison with a simpler alternative (three levels of menus with two options per menu).

Given these advantages, suppliers and automanufacturers have put significant effort into developing speech interfaces for cars. They still have a long way to go. The influential website notes the following [13]:

In the 2012 …, the biggest issue found in today’s vehiclesare the audio, infotainment, and navigation system’s lack of being able to recognize voice commands. This issue was the source of more problems than engine or transmission issues. … Over the four years that the survey questions people on voice recognition systems, problems have skyrocketed 137 percent.

Consumer Reports [14] said the following:

I was feeling pretty good when I spotted that little Microsoft badge on the center console. Now I would be able to access all of those cool SYNC features, right? Wrong.

When I tried to activate text to speech, I was greeted with a dreadful “Not Supported” display. I racked my brain. Did I do something wrong? After all, my phone was equipped seemingly with every feature known to man. … But most importantly, it was powered by Microsoft just like the SYNC system on this 2011 Mustang.

Needing guidance, I went to Ford’s SYNC website. …, I was able to download a 12-page PDF document that listed supported phones. (There is an interactive Sync compatibility guide here, as well.) While I had naively assumed that my high-tech Microsoft phone would work with all the features of “SYNC powered by Microsoft,” the document verified that this was not the case. … Text to speech would only work with a small handful of “dumbphones” that aren’t very popular anymore. Anyone remember the Motorola Razr? That phone was pretty cool a couple of years ago.

One consumer, in commenting about the Chrysler UConnect system said the following [15]:

I have a problem with Uconnect telephone. I input my voice tags but when I then say “Call Mary” the system either defaults to my ’phone book folder or I get 4–6 names on the screen and am asked to “select a line”. I should just say “call Mary home” then I should here my voice with calling “Mary home is that correct”. Can you assist?

Thus, it is appropriate to ask what is known now about the design and evaluation of speech interfaces for cars and how they can be improved. Most engineered systems rely on models, equations, and data to predict system performance and evaluate system alternatives early in development. They do not exist for speech interfaces. Thus, for speech interfaces, the emphasis has been on usability testing, often conducted when development is nearly complete and changes are costly to make.

To be more specific, this paper summarizes the state of the art relating to speech interface design in general, as well as a particular simulation model, namely, one to predict user performance when interacting with a speech interface for destination entry and music selection. The model to be developed will allow for exploration of multiple alternative architectures, recognition rates, and command sets, matters that are very expensive to explore experimentally.

The questions addressed in this paper are as follows.(1)What are some examples of automotive speech interfaces?(2)Who uses speech interfaces, for what, and how often?(3)What are the key research results of the user performance using speech interfaces compared with the user performance using visual-manual interfaces?(4)How should speech interfaces be designed? What are the key design standards and references, design principles, and results from research?(5)How should speech interfaces be assessed and what should be measured?

To answer these questions, both the human factors literature and the linguistics terms and principles are summarized.

2. What Are Some Examples of Automotive Speech Interfaces?

In the USA, current speech interfaces include Ford SYNC, Chrysler UConnect, GM MyLink, Hyundai Genesis, and Toyota navigation with Entune. The commonly supported applications are navigation (e.g., destination entry, route guidance, and traffic information) and music selection (selecting, playing, and pausing songs on MP3 players, AM/FM/XM radios), as well as those related to cellular phones (answering and placing calls, searching contact lists, and various tasks associated with text messages).

To help the reader understand these applications, some examples follow. The extent to which the user and machine are verbose varies between examples.:  Command please.:  Find address.:  Find address. The city please. :  Michigan City.:  Please select the respective line or start spelling.: M I C H I G A N.:  Spell the name in blocks of letters or enter your destination again with change state or enter city.:  Change state.:  Please enter the state name.:  Indiana.:  Please select the respective line or start spelling.:  Line one.:  Line one. Indiana. The city please.:  Michigan City.:  Michigan City. What is the street? :  Wabash Street.:  Wabash Street. And house number or if you do not know that please say show map or start guidance.:  Six zero one.:  Six zero one. Show map or start guidance.:  Start guidance.:  Start guidance. [4]:  Can you find me a Chinese restaurant in Palo Alto with good service?:  There are 11 Chinese restaurants in Palo Alto that have good service. Here are the first few: Little Garden Restaurant China Delight Mandarin Gourmet Restaurant. Should I continue?:  Hmm…Tell me more about, Garden Restaurant?:  Little Garden Restaurant is an expensive Chinese Restaurant. Dress Code is casual, service is good and rating is three stars. It is on El Camino Real in Palo Alto.:  Switch to navigation.:  Take me to little Garden Restaurant, but stop by Bosch Page Mill Office.:  Please say a command.:  USB.:  USB. Is it correct?:  Yes.:  USB. Please say a command.:  PLAY TRACK “Like a Prayer.”:  Play track “Like a prayer.”:  Main menu.:  Phone.:  Phone. Please say a command.:  Call XXX XXX XXXX.:  Calling XXX XXX XXXX.:  AM one thousand six hundred. [It did not work if the user said “AM Sixteen Hundred” for this interface.]:  AM one thousand six hundred.

These systems were developed based on ideas from a number of predecessor systems (Tables 1 and 2). Notice that the core functions were navigation, music selection, and cellular phone support, and that many of them started out as either university or collaborative research projects involving several partners. In several cases, the result was either a product or ideas that later led to products. Of them, probably SYNC has received most of the attention.

SystemResearch/productNavigationRestaurant finderMusic selectionAudio + CDCar
Cellular phone Text message

CHAT [4, 5]ResearchXXX
CU Move [6]ResearchXX
Ford Model U [7]ResearchXXXX
Linguatronic [8]ProductXXXX
SENECA [9]ResearchXXX
SYNC [10]ProductXXXXX
VOIC [11]ResearchXXXX
Volkswagen [12]ProductXX

SystemFull nameDeveloped by

CHATConversational Helper for Automotive TasksCenter for the Study of Language and Information at Stanford University, Research and Technology Center at Bosch, Electronics Research Lab at Volkswagen of America, and Speech Technology and Research Lab at SRI International [4, 5].
CU MoveColorado University MoveUniversity of Colorado speech group in 1999 [6].
Ford Model UFord [7].
LinguatronicDaimlerChrysler Research and Technology in Ulm, Germany and TEMIC in 1996 [8].
SENECASpeech control modules for Entertainment, Navigation, and communication Equipment in CArsEU-project involving DaimlerChrysler, TEMIC Research and Department of Information Technology, University of Ulm [9].
SYNCFord in collaboration with Microsoft and Nuance [10].
VOICVirtual Intelligent Co-DriverEuropean project funded by five different partners: Robert Bosch GmbH, DaimlerChrysler AG, ITCirst, the University of Southern Denmark, and Phonetic Topographics N.V [11].
VolkswagenVolkswagen [12].

The CHAT system uses an event-based, message-oriented system for the architecture with core modules of Natural Language Understanding (NLU), Dialogue Manager (DM), Content Optimization (CO), Knowledge Management (KM), and Natural Language Generation (NLG). CHAT uses the Nuance 8.5 speech recognition engine with class-based n-grams and dynamic grammars, and Nuance Vocalizer as the Text-to-Speech engine. There are three main applications—navigation, MP3 music player, and restaurant finder—to represent important applications in a vehicle [4, 5]. The example for restaurant finder shown earlier is a CHAT dialog.

CU-Move system is an in-vehicle, naturally spoken dialogue system, which can get real-time navigation and route-planning information [6]. The dialogue system is based on the MIT Galaxy-II Hub architecture with base components from CU-Communication system, which is mixed initiative and event driven. This system automatically retrieves the driving direction through Internet with route provider. The dialogue system uses the CMU Sphinx-II speech recognizer for speech recognition and Phoenix Parser for semantic parsing.

A prototype of a conversation system was implemented on the Ford Model U Concept Vehicle and was first shown in 2003 [7]. This system is used for controlling several noncritical automobile operations using speech recognition and a touch screen. The speech recognizer used in this system was speech2Go with adapted acoustic model and other enhancements to improve the speech accuracy. The dialogue manager was a multimodal version of the ETUDE, described by a recursive transition network. Supported applications were climate control, telephone, navigation, entertainment, and system preferences.

Linguatronic is a speech-based command and control system for telephone, navigation, radio, tape, CD, and other applications. The recognizer used in this device was speaker-independent [8].

SENECA SLDS consists of five units: COMMAND head unit connected via an optical Domestic Digital Bus to the Global System for Mobile Communication module, the CD Changer, and Digital Signal Processing module [9]. The system is a command-based speech control of entertainment (radio and CD), navigation, and cellular phones. The speech recognition technology of SENECA SLDS is based on the standard Linguatronic system using the following methods to match the user speech: spell matcher, Java Speech Grammar Format, voice enrollments (user-trained words), and text enrolments. For the dialogue processing, the ENECA SLDS uses a menu-based Command & Control dialogue strategy, including top-down access for main function and side access for subfunction.

SYNC is a fully integrated, voice-activated in-vehicle communication and entertainment system [10] for Ford, Lincoln, and Mercury vehicles in North America. Using commands in multiple languages, such as English, French or Spanish, drivers can operate navigation, portable digital music players, and Bluetooth-enabled mobile phones. The example for music selection shown earlier is a SYNC dialog.

VICO was a research project that concerned a natural-language dialogue prototype [11]. As the interface did not exist, researchers used the Wizard of Oz method to collect the human-computer interaction data. Here, a human operator, the wizard, was simulated system components—speech recognition, natural language understanding, dialogue modeling, and response generation. The goal of this project was to develop a natural language interface allowing drivers to get time, travel (navigation, tourist attraction, and hotel reservation), car, and traffic information safely while driving.

Volkswagen also developed its own in-vehicle speech system [12]. Detailed information about the architecture and methods used to design the system are not available. Supported applications include navigation and cellular phones.

The best-known nonautomotive natural speech interface is Siri, released by Apple in October 2011. Siri can help users make a phone call, find a business and get directions, schedule reminders and meetings, search the web, and perform other tasks supported by built-in apps on the Apple iPhone 4S and iPhone 5.

Similarly, Google’s Voice Actions supports voice search on Android phones (, retrieved May 14, 2012). This application supports sending text messages and email, writing notes, calling businesses and contacts, listening to music, getting directions, viewing a map, viewing websites, and searching webpages. Both Siri and Voice Actions require off-board processing, which is not the case for most in-vehicle speech interfaces.

3. Who Uses Speech Interfaces, for What, and How Often?

Real-world data on the use of speech applications in motor vehicles is extremely limited. One could assume that anyone who drives is a candidate user, but one might speculate that the most technically savvy are the most likely users.

How often these interfaces are used for various tasks is largely unknown. The authors do not know of any published studies on the frequency of use of automotive speech interfaces by average drivers, though they probably exist.

The most relevant information available is a study by Lo et al. [28] concerning navigation-system use, which primarily concerned visual-manual interfaces. In this study, 30 ordinary drivers and 11 auto experts (mostly engineers employed by Nissan) completed a survey and allowed the authors to download data from their personal navigation systems. Data was collected regarding the purpose of trips (business was most common) and the driver’s familiarity with the destination. Interestingly, navigation systems were used to drive to familiar destinations. Within these two groups, use of speech interfaces was quite limited, with only two of the ordinary drivers and two of the auto experts using speech interfaces. The paper also contains considerable details on the method of address entry (street address being used about half of the time followed by point of interest POI) and other information useful in developing evaluations of navigation systems.

Also relevant is the Winter et al. [29] data on typical utterance patterns for speech interfaces, what drivers would naturally say if unconstrained. Included in that paper is information on the number and types of words in utterances, the frequency of specific words, and other information needed to recognize driver utterances for radio tuning, music selection, phone dialing, and POI and street-address entry. Takeda et al. [30] present related research on in-vehicle corpora, which may be a useful resource to address on who, when, and how often the driver used the speech interfaces.

4. What Are the Key Research Results of the User Performance Using Speech Interfaces Compared with the User Performance Using Visual-Manual Interfaces?

There have been a number of studies on this topic. Readers interested in the research should read Barón and Green [31] and then read more recent studies.

Using Barón and Green [31] as a starting point, studies of the effects of speech interfaces on driving are summarized in four tables. Table 3 summarizes bench-top studies of various in-vehicle speech interfaces. Notice that the value of the statistics varied quite widely between speech interfaces, mainly because the tasks examined were quite different. As an example for CU-Communicator [16], the task required the subject to reserve a one-way or round-trip flight within or outside the United States with a phone call. Performing this task involved many turns between users and machines (total 38 turns) and the task took almost 4.5 minutes to complete. Within speech interfaces, task-completion time varied from task to task depending on the task complexity [11, 12].

SystemCHAT [4]CHAT [5]CU Communicator [16]CU Move [6]SENECA [9]VOIC [11]Volkswagen [12]

Tasks(1) NAV
(2) Restaurant finder (RF)
(1) MP3
(2) Restaurant finder (RF)
Phone for travel planNAV(1) NAV
(2) Phone dialing
(3) Address book
(1) NAV
(2) Current time
(3) Tourism
(4) Fuel
(5) Car manual
(6) Hotel reservation
(7) Traffic information
(1) NAV
(2) Map control
(3) Phone

Completion time (s)260.3Average: 63(1) 73
(2) 15
(3) 146
(4) 78
(5) 57
(6) 180
(7) 53

Completion rate98%MP3: 98%
RF: 94%
73.6%Average: 79%

Turns12.31RF: 4.11User: 19
Machine: 19

Word recognition Accuracy2 (%)NAV: 85.5%
RF: 85%
MP3: 90%
RF: 85%

Word error rate (%)26%30–65%

User satisfaction rating31.98MP3: 2.24
RF: 2.04

Turn is defined as one user utterance to the system during a dialog exchange between the user and the system while attempting to perform the task.
2Word recognition accuracy (WA).
   : total number of words in reference.
  Ws: number of reference words which were substituted in output.
  Wi: number of reference words which were inserted in output.
  Wd: number of reference words which were deleted in output.
3User satisfaction rating: 1 = strong agreement; 5 = strong disagreement.

Table 4, which concerns driving performance, shows that the use of speech interfaces as opposed to visual-manual interfaces led to better lane keeping (e.g., lower standard deviation of lane position).

StudyMethodLane keepingBrake reaction timePeripheral detection timeFollowing distance

Carter and Graham [17]SimulatorS < MS < M
Forlines et al. [18]SimulatorS < MNo diff.
Garay-Vega et al. [19]SimulatorNo diff.
Gärtner et al. [20]On roadS < M
Itoh et al. [21]SimulatorS < MNo diff.
Maciej and Vollrath [22]SimulatorS < M
McCallum et al. [23]SimulatorNo diff.No diff.
Minker et al. [9]On roadS < M
Ranney et al. [24]On roadS < M (0.8 versus 0.87 s)
Shutko et al. [25]SimulatorS < MS < M (Except incoming call)
Tsimhoni et al. [26]SimulatorS < KS < K (88 versus 167 m)
Villing et al. [27]On road

Table 5 shows that task completion times for speech interfaces were sometimes shorter than that for visual-manual interfaces and sometimes longer, even though people speak faster than they can key in responses. This difference is due to the inability of the speech interface to correctly recognize what the driver says, requiring utterances to be repeated. Speech recognition accuracy was an important factor that affected the task performance. Kun et al. [33] reported that low recognition accuracy (44%) can lead to greater steering angle variance. Gellatly and Dingus [34] reported that driving performance (peak lateral acceleration and peak longitudinal acceleration) was not statistically affected until the 60% recognition accuracy level was reached. Gellatly and Dingus [34] also showed that the task completion time was also affected when the speech recognition accuracy was lower than 90%. Although speech recognition accuracy was found to affect driving and task performance, no research has been reported on drivers’ responses to errors, how long drivers need to take to correct errors, or what strategies drivers use to correct errors. Understanding how users interact with the spoken dialogue systems can help designers improve system performance and make drivers feel more comfortable using speech interfaces.

StudyTask completion timeSpeech recognizer rateTask completion rate

Carter and Graham [17]S > M92.7%
Forlines et al. [18]S < M (18.0 versus 25.2 s)
Garay-Vega et al. [19] S (dialog-based) > M
S (query-based) < M
Gärtner et al. [20] S > M
Simple: 24.6 versus 12.8 s
Complex: 74.4 versus 58.7 s
79.4% (recognition error rate: 20.6%)
Minker et al. [9]S < M (63 versus 84 sec)S < M (79 versus 90%)
Ranney et al. [24]No difference
Shutko et al. [25]S < M (except dialing phone)
Villing et al. [27]S > M

Table 6 shows that when using speech interfaces while driving, as opposed to visual-manual interfaces, subjective workload was less, fewer glances were required, and glance durations were shorter.

StudySubjective ratingDriving behavior—glances
Workload PreferenceGlance durationNumber of glances

Carter and Graham [17]S < M
Faerber and Meier-Arendt [32]S is preferredRadio or CD control:
M: 1 s; S: 1/3 s
Using phone: M: 1 s
Radio or CD control:
M: 3; S: 1.1
Using phone: M: 12; S: 0
Garay-Vega et al. [19]S (query-based) < MS < M
Gärtner et al. [20]No differenceSimple task: no difference
Complex tasks: S < M
Itoh et al. [21]S < MS < M
Maciej and Vollrath [22]S < M
McCallum et al. [23]S < M
Minker et al. [9]S is preferred
Shutko et al. [25]S < M
(Except receiving an incoming call)

In general, driving performance while using speech interfaces is generally better than when using visual-manual interfaces. That is, speech interfaces are less distracting.

5. How Should Speech Interfaces Be Designed? What Are the Key Design Standards and References, Design Principles, and Results from Research?

5.1. Relevant Design and Evaluation Standards

For speech interfaces, the classic design guidelines are that of Schumacher et al. [35], and the one set that is not very well known, but extremely useful, is the Intuity guidelines [36]. Najjar et al. [37] described user-interface design guidelines for speech recognition applications. Hua and Ng [38] also proposed guidelines on in-vehicle speech interfaces based on a case study.

Several technical standards address the topic of the evaluation of speech system performance. These standards, such as ISO 9921: 2003 (Ergonomics—Assessment of speech communication), ISO 19358: 2002 (Ergonomics—Construction and application of tests for speech technology), ISO/IEC 2382-29: 1999 (Artificial intelligence—Speech recognition and synthesis), and ISO 8253-3: 2012 (Acoustics—Audiometric tests methods—Part 3: Speech Audiometry), focus on the evaluation of the whole system and its components [3942]. However, no usability standards related to speech interfaces have emerged other than ISO/TR 16982: 2002 (Ergonomics of human-system interaction—Usability methods supporting human-centered design) [43].

From its title (Road vehicles—Ergonomic aspects of transport information and control systems—Specifications for in-vehicle auditory presentation), one would think that ISO 15006: 2011 [44] is relevant. In fact, ISO 15006 concerns nonspoken warnings.

There are standards in development. SAE J2988, Voice User Interface Principles and Guidelines [45], contains 19 high-level principles (e.g., principle 17: “Audible lists should be limited in length and content so as not to overwhelm the user’s short-term memory.”). Unfortunately, no quantitative specifications are provided. The draft mixes definitions and guidance in multiple sections making the document difficult to use, does not support guidance with references, and, in fact, has no references.

The National Highway Traffic Safety Administration (NHTSA) of the US Department of Transportation posted proposed visual-manual driver-distraction guidelines for in-vehicle electronic devices for public comment on February 15, 2012 ('Distraction'+Guidelines+for+Automakers, retrieved May 15, 2012). NHTSA has plans for guidelines for speech interfaces.

The distraction focus group of the International Telecommunication Union (FG-Distraction-ITU) is interested in speech interfaces and may eventually issue documents on this topic, but what and when are unknown. In addition, various ITU documents that concern speech-quality assessment may be relevant, though they were intended for telephone applications. ITU-P.800 (methods for subjective determination of transmission quality) and related documents are of particular interest. See

5.2. Key Books

There are a number of books on speech interface design, with the primary references being Hopper’s classic [46], Balentine and Morgan [47], Cohen et al. [48], and Harris [49]. A more recent reference is Lewis [50].

5.3. Key Linguistic Principles

The linguistic literature provides a framework for describing the interaction, the kinds of errors that occur, and how they could be corrected. Four topics are touched upon here.

5.3.1. Turn and Turn-Taking

When can the user speak? When does the user expect the system to speak? Taking a turn refers to an uninterrupted speech sequence. Thus, the back-and-forth dialog between a person and a device is turn-taking, and the number of turns is a key measure of an interface’s usability, with fewer turns indicating a better interface. In general, overlapping turns, where both parties speak at the same time, account for less than 5% of the turns that occur while talking [51]. The amount of time between turns is quite small, generally less than a few hundred milliseconds. Given the time required to plan an utterance, planning starts before the previous speaker finishes the utterance.

One of the important differences between human-human and human-machine interactions is that humans often provide nonverbal feedback that indicates whether they understand what is said (e.g., head nodding), which facilitates interaction and control of turn-taking. Most speech interfaces do not have the ability to process or provide this type of feedback.

A related point is that most human-human interactions accept interruptions (also known as barge-in), which makes interactions more efficient and alters turn taking. Many speech interfaces do support barge-in, which requires the users to press the voice-activation button. However, less than 10% of subjects (unpublished data from the authors) knew and used this function.

5.3.2. Utterance Types (Speech Acts)

Speech acts refer to the kinds of utterances made and their effect [53]. According to Akmajian et al. [54], there are four categories of speech acts.(i)Utterance acts include uttering sounds, syllables, words, phrases, and sentences from a language including filler words (“umm”).(ii)Illocutionary acts include asking, promising, answering, and reporting. Most of what is said in a typical conversation is this type of act.(iii)Perlocutionary acts are utterances that produce an effect on the listener, such as inspiration and persuasion.(iv)Propositional acts are acts in which the speaker refers to or predicts something.

Searle [55] classifies speech acts into five categories.(i)Assertives commit the speaker to address something (suggesting, swearing, and concluding).(ii)Directives get the listener to do something (asking, ordering, inviting).(iii)Commissives commit the speaker to some future course of action (promising, planning).(iv)Expressives express the psychological state of the speaker (thanking, apologizing, welcoming).(v)Declarations bring a different state to either speaker or listener (such as “You are fired”).

5.3.3. Intent and Common Understanding (Conversational Implicatures and Grounding)

Sometimes speakers can communicate more than what is uttered. Grice [56] proposed that conversations are governed by the cooperative principle, which means that speakers make conversational contributions at each turn to achieve the purpose or direction of a conversation. He proposed four high levels conversational maxims that may be thought of as usability principles (Table 7).


Maxim of Quantity:
be informative.
Machine: Please say the street name.
User: 2901 Baxter Road (2901 is the house number)
(i) Make your contribution of information as is required, that is, for the current purpose of the conversation.
(ii) Do not make your contribution more informative than is required.

Maxim of Quality:
make your contribution one that is true:
U: “Toledo Zoo, Michigan” (but Toledo is in Ohio)(i) Do not say what you believe to be false.
(ii) Do not say that for which you lack evidence.

Maxim of Relevance:
be relevant.
U: “I want to go to “Best Buy”” and the system responds with all Best Buy stores, including those hundreds of miles away, not just the local ones.

Maxim of Manner:
be perspicuous
(i) M: Please say set as destination, dial or back.
   U: Dial. O no Don’t dial, Back (User want to say “back”)
(ii) M: Please say the POI category.
    U: Let’s see. Recreation
(i) Avoid obscurity of expression
(ii) Avoid unnecessary ambiguity
(iii) Be brief (avoid unnecessary prolixity)
(iv) Be orderly.

5.3.4. Which Kinds of Errors Can Occur?

Skanztze [52] provides one of the best-known schemes for classifying errors (Table 8). Notice that Skanztze does so from the perspective of a device presenting an utterance and then processing a response from a user.

ModulesPossible sources of errors

Speech detectionTruncated utterances, artifacts such as noise and side talk; barge-in problems
Speech recognitionInsertions, deletions, substitutions
Language processing/parsingConcept failure, speech act tagging
Dialogue managerError in reference resolution, error in plan recognition
Response generationAmbiguous references, too much information presented at once, TTS quality, audio quality

Véronis [57] presents a more detailed error-classification scheme that considers device and user errors, as well as the linguistic level (lexical, syntactic, semantic). Table 9 is an enhanced version of that scheme. Competence, one of the characteristics in his scheme, is the knowledge the user has of his or her language, whereas performance is the actual use of the language in real-life situations [58]. Competence errors result from the failure to abide by linguistic rules or from a lack of knowledge of those rules (“the information from users is not in the database”), whereas performance errors are made despite the knowledge of rules (“the interface does not hear users’ input correctly”).

System User

Lexical level
Letter substitution
Letter insertion
Letter deletion
Word missing in dictionary
Missing inflection rule
Letter substitution
Letter insertion
Letter deletion
Letter transposition
Syllabic error
Slips of tongue
Nonword or completely garbled word

Syntactic level
(sentence structure)
Missing ruleWord substitution
Word insertion
Word deletion
Word transposition
Construction error

Semantic level
Incomplete or contradictory knowledge representation
Unexpected situation
Conceptual error includes:
incomplete or contradictory knowledge representation 
Pragmatic error includes:
dialogue law violation

As an example, a POI category requested by the user that was not in the database would be a semantic competence error. Problems in spelling a word would be a lexical performance error. Inserting an extra word in a sequence (“iPod iPod play …”) would be a lexical performance error.

A well-designed speech interface should help avoiding errors, and, when they occur, facilitating correction. Strategies to correct errors include repeating and rephrasing the utterances, spelling out words, contradicting a system response, correcting using a different modality (e.g., manual entry instead of speech), and restarting, among others [5962].

Knowing how often these strategies occur suggests what needs to be supported by the interface. The SENECA project [9, 20] revealed that the most frequent errors for navigation tasks were spelling problems of various types, entering or choosing the wrong street, and using wrong commands. For phone dialing tasks, the most frequent errors were stops within digit sequences. In general, most of the user errors were vocabulary errors (partly spelling errors), dialogue flow errors, and PTA (push to active) errors, that is, missing or inappropriate PTA activation.

Lo et al. [63] reported that construction and relationship errors were 16% and 37%, respectively. Construction errors occur when subjects repeat words, forget to say command words (a violation of grounding), or forget to say any other words that were given. Relationship errors occur when subjects make incorrect matches between the given words and song title, album name, and/or artist name. Relationship errors were common because subjects were not familiar with the given songs/albums/artists.

6. How Should Speech Interfaces Be Assessed and What Should Be Measured?

6.1. What Methods Should Be Used?

Given the lack of models to predict user performance with speech interfaces, the evaluation of the safety and usability (usability testing) of those interfaces has become even more important. Evaluations may either be performed only with the system itself (on a bench top) or with the system integrated into a motor vehicle (or a simulator cab) while driving.

The most commonly used method to evaluate in-vehicle speech interfaces is the Wizard of Oz method [4, 5, 11, 16, 6466], sometimes implemented using Suede [67]. In a Wizard of Oz experiment, subjects believe that they are interacting with a computer system, not a person simulating one. The “wizard” (experimenter), who is remote from the subject, observes the subject’s actions and simulates the system’s responses in real-time. To simulate a speech-recognition application, the wizard would type what users say, or in a text-to-speech system, they read the text output, often in a machine-like voice. Usually, it is much easier to tell a person how to emulate a machine than to write the software to tell a computer to do it. The Wizard of Oz method allows for the rapid simulation of speech interfaces and the collection of data from users interacting with a speech interface, allowing for multiple iterations of the interface to be tested and redesigned.

6.2. What Should Be Measured?

Dybkjaer has written several papers on speech interface evaluation, the most thorough of which is Dybkjær et al. [68]. That paper identified a number of variables that could be measured (Table 10), in part because there are many attributes to consider.


Whole systemTask completion time, task completion rate, transaction success, number of interaction problems, query density, concept efficiency
Speech recognitionWord and sentence error rate, vocabulary coverage, perplexity
Speech synthesizerUser perception, speech intelligibility, pleasantness, naturalness
Language understandingLexical coverage, grammar coverage, real-time performance, concept accuracy, concept error rate

Walker et al. [69] proposed a framework of usability evaluation of spoken dialogue systems, known as PARADISE (PARAdigm for DIalogue System Evaluation). (See [70] for criticisms.) Equations were developed to predict dialog efficiency (which depends on mean elapsed time and the mean number of user moves) and dialog quality costs (which depends on the number of missing responses, the number of errors, and many other factors, and task success, measured by the Kappa coefficient and defined below): where = proportion of times that the actual set of dialogues agree with scenario keys; = proportion of times that the dialogues and the keys are expected to agree by chance.

In terms of performance while driving, there is no standard or common method for evaluating speech interfaces, with evidence from bench-top, simulator, and on-road experiments being used. There are two important points to keep in mind when conducting such evaluations. First, in simulator and on-road experiments, the performance on the secondary speech interface task depends on the demand of the primary driving task. However, the demand or workload of that task is rarely quantified [71, 72]. Second, there is great inconsistency in how secondary-task performance measures are defined, if they are defined at all, making the comparison of evaluations quite difficult [73]. (See [74] for more information.) Using the definitions in SAE Recommended Practice J2944 [75] is recommended.

7. Summary

The issues discussed in this paper are probably just a few of those which should be considered in a systematic approach to the design and development of speech interfaces.

7.1. What Are Some Examples of Automotive Speech Interfaces?

Common automotive examples include CHAT, CU Move, Ford Model U, Linguatronic, SENECA, SYNC, VOIC, and Volkswagen. Many of these examples began as collaborative projects that eventually became products. SYNC is the best known.

Also important are nonautomotive-specific interfaces that will see in-vehicle use, in particular, Apple Siri for the iPhone and Google Voice Actions for Android phones.

7.2. Who Uses Speech Interfaces, for What, and How Often?

Unfortunately, published data on who uses speech interfaces and how real drivers in real vehicles use them is almost zero. There are several studies that examine how these systems are used in driving simulators, but those data do not address this question.

7.3. What Are the Key Research Results of the User Performance Using Speech Interfaces Compared with the User Performance Using Visual Manual Interfaces?

To understand the underlying research, Barón and Green’s study [31] is a recommended summary. Due to the difference of task complexity while testing, comparing alternative speech systems is not so easy. However, when compared with visual-manual interfaces, speech interfaces led to consistently better lane keeping, shorter peripheral detection time, lower workload ratings, and shorter glance durations away from the road. Task completion time was sometimes greater and sometimes less, depending upon the study.

7.4. How Should Speech Interfaces Be Designed? What Are the Key Design Standards and References, Design Principles, and Results from Research?

There are a large number of relevant technical standards to help guide speech interfaces. In terms of standards, various ISO standards (e.g., ISO 9921, ISO 19358, ISO 8253) focus on the assessment of the speech interaction, not on design. Speech-quality assessment is considered by ITU-P.800. For design, key guidelines include [3538]. A number of books also provide useful design guidance including [4650].

Finally, the authors would recommend that any individual seriously engaged in speech-interface design should understand the linguistic terms and principles (turns, speech acts, grounding, etc.) as the literature provides several useful frameworks for classifying errors and information that provides clues as to how to reduce errors associated with using a speech interface.

7.5. How Should Speech Interfaces Be Assessed and What Should Be Measured?

The Wizard of Oz method is commonly used in the early stages of interface development. In that method, an unseen experimenter behind the scenes simulates the behavior of a speech interface by recognizing what the user says or is speaking in response to what the user says, or both. Wizard of Oz simulations take much less time to implement than other methods.

As automotive speech interfaces move close to production, the safety and usability of those interfaces are usually assessed in a driving simulator, and sometimes on the road. The linguistics literature provides a long list of potential measures of the speech interface that could be used, with task time being the most important. Driving-performance measures, such as standard deviation of lane position and gap variability, are measured as eyes-off-the-road time. These studies often have two key weaknesses: (1) the demand/workload of the primary task is not quantified, yet performance on the secondary speech task can depend on its demand and (2) measures and statistics describing primary task performance are not defined. A solution to the first problem is to use equations being developed by the second author to quantify primary task workload. The solution to the second problem is to use the measures and statistics in SAE Recommended Practice J2944 [75] and refer to it.

Driver distraction is and continues to be a major concern. Some view speech interfaces as a distraction-reducing alternative to visual-manual interfaces. Unfortunately, at this point, actual use by drivers and data on that use is almost zero. There is some information on how to test speech interfaces, but technical standards cover only a limited number of aspects.

There is very little to support design other than guidelines. For most engineered systems, developers use equations and models to predict system and user performance, with testing serving as verification of the design. For speech interfaces, those models do not exist. This paper provides some of the background information needed to create those models.


  1. J. C. Stutts, D. W. Reinfurt, and L. Staplin, “The role of driver distraction in traffic crashes,” AAA Foundation for Traffic Safety, Washington, DC, USA, 2001, View at: Google Scholar
  2. O. Tsimhoni, D. Smith, and P. Green, “Address entry while driving: speech recognition versus a touch-screen keyboard,” Human Factors, vol. 46, no. 4, pp. 600–610, 2004. View at: Publisher Site | Google Scholar
  3. J. D. Lee, B. Caven, S. Haake, and T. L. Brown, “Speech-based interaction with in-vehicle computers: the effect of speech-based e-mail on drivers' attention to the roadway,” Human Factors, vol. 43, no. 4, pp. 631–640, 2001. View at: Google Scholar
  4. F. Weng, B. Yan, Z. Feng et al., “CHAT to your destination,” in Proceedings of the 8th SIGdial Workshop on Discourse and Dialogue, pp. 79–86, Antwerp, Belgium, 2007. View at: Google Scholar
  5. F. Weng, S. Varges, B. Raghunathan et al., “CHAT: a conversational helper for automotive tasks,” in Proceedings of the 9th International Conference on Spoken Language Processing (Inter-Speech/ICSLP '06), pp. 1061–1064, Pittsburgh, Pa, USA, September 2006. View at: Google Scholar
  6. J. H. L. Hansen, J. Plucienkowski, S. Gallant, B. Pellom, and W. Ward, “‘CU-Move’: robust speech processing for in-vehicle speech system,” in Proceedings of the International Conference on Spoken Language Processing (ICSLP '00), vol. 1, pp. 524–527, Beijing, China, 2000. View at: Google Scholar
  7. R. Pieraccini, K. Dayanidhi, J. Bloom et al., “Multimodal conversational systems for automobiles,” Communications of the ACM, vol. 47, no. 1, pp. 47–49, 2004. View at: Publisher Site | Google Scholar
  8. P. Heisterkamp, “Linguatronic product-level speech system for Mercedes-Benz cars,” in Proceedings of the 1st International Conference on Human Language Technology Research, pp. 1–2, Association for Computational Linguistics, San Diego, Calif, USA, 2001. View at: Google Scholar
  9. W. Minker, U. Haiber, P. Heisterkamp, and S. Scheible, “The SENECA spoken language dialogue system,” Speech Communication, vol. 43, no. 1-2, pp. 89–102, 2004. View at: Publisher Site | Google Scholar
  10. Sync,
  11. P. Geutner, F. Steffens, and D. Manstetten, “Design of the VICO spoken dialogue system: evaluation of user expectations by wizard-of-oz experiments,” in Proceedings of the 3rd International Conference on Language Resources and Evaluation (LREC '02), Las Palmas, Spain, 2002. View at: Google Scholar
  12. J. C. Chang, A. Lien, B. Lathrop, and H. Hees, “Usability evaluation of a Volkswagen Group in-vehicle speech system,” in Proceedings of the 1st International Conference on Automotive User Interfaces and Interactive Vehicular Applications (ACM '09), pp. 137–144, Essen, Germany, September 2009. View at: Publisher Site | Google Scholar
  13. T. Dorchies, “Come again? Vehicle voice recognition biggest problem in J.D. Power and Associates study,” View at: Google Scholar
  14. “Consumer Reports, The Ford SYNC system read test messages…in theory,” View at: Google Scholar
  15. JustAnswer, “Chrysler 300c I have a problems with Uconnect telephone,” View at: Google Scholar
  16. B. Pellom, W. Ward, J. Hansen et al., “University of Colorado dialog systems for travel and navigation,” in Proceedings of the 1st International Conference on Human language Technology Research, Association for Computational Linguistics, San Diego, Calif, USA, 2001. View at: Google Scholar
  17. C. Carter and R. Graham, “Experimental comparison of manual and voice controls for the operation of in-vehicle systems,” in Proceedings of the 14th Triennial Congress of the International Ergonomics Association and 44th Annual Meeting of the Human Factors and Ergonomics Association (IEA/HFES '00), vol. 44, pp. 286–289, Human Factors and Ergonomics Society, Santa Monica, CA, USA, August 2000. View at: Google Scholar
  18. C. Forlines, B. Schmidt-Nielsen, B. Raj, K. Wittenburg, and P. Wolf, “A comparison between spoken queries and menu-based interfaces for in-car digital music selection,” in Proceedings of the International Conference on Human-Computer Interaction (INTERACT '05), pp. 536–549, Rome, Italy, 2005. View at: Google Scholar
  19. L. Garay-Vega, A. K. Pradhan, G. Weinberg et al., “Evaluation of different speech and touch interfaces to in-vehicle music retrieval systems,” Accident Analysis and Prevention, vol. 42, no. 3, pp. 913–920, 2010. View at: Publisher Site | Google Scholar
  20. U. Gärtner, W. König, and T. Wittig, “Evaluation of manual vs. speech input when using a driver information system in real traffic,” in Proceedings of the Driving Assessment 2001: The First International Driving Symposium on Human Factors in Driving Assessment, Training and Vehicle Design, Aspen, Colo, USA, 2001. View at: Google Scholar
  21. K. Itoh, Y. Miki, N. Yoshitsugu, N. Kubo, and S. Mashimo, “Evaluation of a voice-activated system using a driving simulator,” SAE World Congress & Exhibition, SAE Tech 2004-01-0232, Society of Automotive Engineers, Warrendale, Pa, USA, 2004. View at: Google Scholar
  22. J. Maciej and M. Vollrath, “Comparison of manual vs. speech-based interaction with in-vehicle information systems,” Accident Analysis and Prevention, vol. 41, no. 5, pp. 924–930, 2009. View at: Publisher Site | Google Scholar
  23. M. C. McCallum, J. L. Campbell, J. B. Richman, J. L. Brown, and E. Wiese, “Speech recognition and in-vehicle telematics devices: potential reductions in driver distraction,” International Journal of Speech Technology, vol. 7, no. 1, pp. 25–33, 2004. View at: Publisher Site | Google Scholar
  24. T. A. Ranney, J. L. Harbluk, and Y. I. Noy, “Effects of voice technology on test track driving performance: implications for driver distraction,” Human Factors, vol. 47, no. 2, pp. 439–454, 2005. View at: Publisher Site | Google Scholar
  25. J. Shutko, K. Mayer, E. Laansoo, and L. Tijerina, “Driver workload effects of cell phone, music player, and text messaging tasks with the Ford SYNC voice interface versus handheld visual-manual interfaces,” SAE World Congress & Exhibition, SAE Tech 2009-01-0786, Society of Automotive Engineers, Warrendale, Pa, USA, 2009. View at: Google Scholar
  26. O. Tsimhoni, D. Smith, and P. Green, “Destination entry while driving: speech recognition versus a touch-screen keyboard,” Tech. Rep. UMTRI-2001-24, University of Michigan Transportation Research Institute, Ann Arbor, Mich, USA, 2002. View at: Google Scholar
  27. J. Villing, C. Holtelius, S. Larsson, A. Lindstrom, A. Seward, and N. Aberg, “Interruption, resumption and domain switching in in-vehicle dialogue,” in Proceedings of the 6th International Conference on Natural Language Processing, pp. 488–499, Gothenburg, Sweden, 2008. View at: Google Scholar
  28. V. E.-W. Lo, P. A. Green, and A. Franzblau, “Where do people drive? Navigation system use by typical drivers and auto experts,” Journal of Navigation, vol. 64, no. 2, pp. 357–373, 2011. View at: Publisher Site | Google Scholar
  29. U. Winter, T. J. Grost, and O. Tsimhoni, “Language pattern analysis for automotive natural language speech applications,” in Proceedings of the 2nd International Conference on Automotive User Interfaces and Interactive Vehicular Applications (ACM '10), pp. 34–41, Pittsburgh, Pa, USA, November 2010. View at: Publisher Site | Google Scholar
  30. K. Takeda, J. H. L. Hensen, P. Boyraz, L. Malta, C. Miyajima, and H. Abut, “International large-scale vehicle corpora for research on driver behavior on the road,” IEEE Transactions on Intelligent Transportation Systems, vol. 12, pp. 1609–1623, 2011. View at: Google Scholar
  31. A. Barón and P. A. Green, “Safety and usability of speech interfaces for in-vehicle tasks while driving: a brief literature review,” Tech. Rep. UMTRI-2006-5, University of Michigan Transportation Research Institute, Ann Arbor, Mich, USA, 2006. View at: Google Scholar
  32. B. Faerber and G. Meier-Arendt, “Speech control systems for handling of route guidance, radio and telephone in cars: results of a field experiment,” in Vision in Vehicle—VII, A. G. Gale, Ed., pp. 507–515, Elsevier, Amsterdam, The Netherlands, 1999. View at: Google Scholar
  33. A. Kun, T. Paek, and Z. Medenica, “The effect of speech interface accuracy on driving performance,” in Proceedings of the 8th Annual Conference of the International Speech Communication Association (Interspeech '07), pp. 1326–1329, Antwerp, Belgium, August 2007. View at: Google Scholar
  34. A. W. Gellatly and T. A. Dingus, “Speech recognition and automotive applications: using speech to perform in-vehicle tasks,” in Proceedings of the Human Factors and Ergonomics Society 42nd Annual Meeting, pp. 1247–1251, Santa Monica, Calif, USA, October 1998. View at: Google Scholar
  35. R. M. Schumacher, M. L. Hardzinski, and A. L. Schwartz, “Increasing the usability of interactive voice response systems: research and guidelines for phone-based interfaces,” Human Factors, vol. 37, no. 2, pp. 251–264, 1995. View at: Publisher Site | Google Scholar
  36. Intuity Conversant Voice Information System Version 5. 0 Application Design Handbook, AT&T Product Documentation Development, Denver, Colo, USA, 1994.
  37. L. J. Najjar, J. J. Ockeman, and J. C. Thompson, “User interface design guidelines for speech recognition applications,” presented at IEEE VARIS 98 Workshop, Atlanta, Ga, USA, 1998, View at: Google Scholar
  38. Z. Hua and W. L. Ng, “Speech recognition interface design for in-vehicle system,” in Proceedings of the 2nd International Conference on Automotive User Interfaces and Interactive Vehicular Applications, pp. 29–33, ACM, Pittsburgh, Pa, USA, 2010. View at: Google Scholar
  39. “Ergonomics—Assessment of Speech Communication,” ISO Standard 9921, 2003. View at: Google Scholar
  40. “Ergonomics—Construction and Application of Tests for Speech Technology,” Tech. Rep. ISO/TR 19358, 2002. View at: Google Scholar
  41. “Information Technology—Vocabulary—Part 29: Artificial Intelligence—Speech Recognition and Synthesis,” ISO/IEC Standard 2382-29, 1999. View at: Google Scholar
  42. “Acoustics—Audiometric Test Methods—Part 3: Speech Audiometry,” ISO Standard 8253-3, 2012. View at: Google Scholar
  43. “Ergonomics of human-system interaction—Usability methods supporting human-centered design,” ISO Standard 16982, 2002. View at: Google Scholar
  44. “Road vehicles—Ergonomic aspects of transport information and control systems—Specifications for in-vehicle auditory presentation,” ISO Standard 15006, 2011. View at: Google Scholar
  45. Voice User Interface Principles and Guidelines (Draft), SAE Recommended Practice J2988, 2012.
  46. R. Hopper, Telephone Conversation, Indiana University Press, Bloomington, IN, USA, 1992.
  47. B. Balentine and D. P. Morgan, How to Build A Speech Recognition Application, Enterprise Integration Group, San Ramon, Calif, USA, 1999.
  48. M. H. Cohen, J. P. Giangola, and J. Balogh, Voice Interface Design, Pearson, Boston, Mass, USA, 2004.
  49. R. A. Harris, Voice Interaction Design, Morgan Kaufmann, San Francisco, Calif, USA, 2005.
  50. J. R. Lewis, Practical Speech User Interface Design, CRC Press, Boca Raton, Fla, USA, 2011.
  51. S. C. Levinson, Pragmatics, Cambridge University Press, New York, NY, USA, 1983.
  52. G. Skanztze, “Error detection in spoken dialogue systems,” 2002, View at: Google Scholar
  53. J. L. Austin, How To Do Things With Words, Harvard University Press, Cambridge, Mass, USA, 1962.
  54. A. Akmajian, R. A. Demers, A. K. Farmer, and R. M. Harnish, Linguistics: An Introduction To Language and Communication, MIT Press, Cambridge, Mass, USA, 5th edition, 2001.
  55. J. R. Searle, “A taxonomy of illocutionary,” in Language, Mind and Knowledge, Minnesota Studies in the Philosophy of Science, K. Gunderson, Ed., vol. 7, pp. 344–369, 1975. View at: Google Scholar
  56. H. P. Grice, “Logic and conversation,” in Syntax and Semantics 3: Speech Acts, P. Coole and J. L. Morgan, Eds., pp. 41–58, Academic Press, New York, NY, USA, 1975. View at: Google Scholar
  57. J. Véronis, “Error in natural language dialogue between man and machine,” International Journal of Man-Machine Studies, vol. 35, no. 2, pp. 187–217, 1991. View at: Google Scholar
  58. N. Chomsky, Aspects of Theory of Syntax, The MIT Press, Cambridge, Mass, USA, 1965.
  59. M. L. Bourguet, “Towards a taxonomy of error-handling strategies in recognition-based multi-modal human-computer interfaces,” Signal Processing, vol. 86, no. 12, pp. 3625–3643, 2006. View at: Publisher Site | Google Scholar
  60. C.-M. Karat, C. Halverson, D. Horn, and J. Karat, “Patterns of entry and correction in large vocabulary continuous speech recognition systems,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 568–575, ACM, Pittsburgh, Pa, USA, May 1999. View at: Google Scholar
  61. K. Larson and D. Mowatt, “Speech error correction: the story of the alternates list,” International Journal of Speech Technology, vol. 6, no. 2, pp. 183–194, 2003. View at: Publisher Site | Google Scholar
  62. D. Litman, M. Swerts, and J. Hirschberg, “Characterizing and predicting corrections in spoken dialogue systems,” Computational Linguistics, vol. 32, no. 3, pp. 417–438, 2006. View at: Publisher Site | Google Scholar
  63. E.-W. Lo, S. M. Walls, and P. A. Green, “Simulation of iPod music selection by drivers: typical user task time and patterns for manual and speech interfaces,” Tech. Rep. UMTRI-2007-9, University of Michigan Transportation Research Institute, Ann Arbor, Mich, USA, 2007. View at: Google Scholar
  64. J. F. Kelley, “An empirical methodology for writing user-friendly natural language computer applications,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 193–196, ACM, Boston, Mass, USA, 1983. View at: Google Scholar
  65. J. D. Gould, J. Conti, and T. Hovanyecz, “Composing letters with a simulated listening typewriter,” Communications of the ACM, vol. 26, no. 4, pp. 295–308, 1983. View at: Publisher Site | Google Scholar
  66. P. Green and L. Wei-Hass, “The Wizard of Oz: a tool for repaid development of user interfaces,” Tech. Rep. UMTRI-1985-27, University of Michigan Transportation Research Institute, Ann Arbor, Mich, USA, 1985. View at: Google Scholar
  67. A. K. Sinah, S. R. Klemmer, J. Chen, J. A. Landay, and C. Chen, “Suede: iterative, informal prototyping for speech interfaces,” in Proceedings of the CHI 2001 Proceedings, Association for Computing Machinery, New York, NY, USA, 2011. View at: Google Scholar
  68. L. Dybkjær, N. O. Bernsen, and W. Minker, “Evaluation and usability of multimodal spoken language dialogue systems,” Speech Communication, vol. 43, no. 1-2, pp. 33–54, 2004. View at: Publisher Site | Google Scholar
  69. M. Walker, C. Kamm, and D. Litman, “Towards developing general models of usability with PARADISE,” Natural Language Engineering, vol. 6, pp. 363–377, 2000. View at: Google Scholar
  70. M. Hajdinjak and F. Mihelič, “The PARADISE evaluation framework: issues and findings,” Computational Linguistics, vol. 32, no. 2, pp. 263–272, 2006. View at: Publisher Site | Google Scholar
  71. J. Schweitzer and P. A. Green, “Task acceptability and workload of driving urban roads, highways, and expressway: ratings from video clips,” Tech. Rep. UMTRI-2006-6, University of Michigan Transportation Research Institute, Ann Arbor, Mich, USA, 2007. View at: Google Scholar
  72. P. Green, B. T.-W. Lin, J. Schweitzer, H. Ho, and K. Stone, “Evaluation of a method to estimate driving workload in real time: watching clips versus simulated driving,” Tech. Rep. UMTRI-2011-29, University of Michigan Transportation Research Institute, Ann Arbor, Mich, USA, 2011. View at: Google Scholar
  73. P. Green, “Using standards to improve the replicability and applicability of driver interface research,” in Proceedings of the 4th International Conference on Automotive User Interfaces and Interactive Vehicular (AutomotiveUI '12), Portsmouth, UK. View at: Google Scholar
  74. M. R. Savino, Standardized names and definitions for driving performance measures [Ph.D. thesis], Department of Mechanical Engineering, Tufts University, Medford, Ore, USA, 2009.
  75. Operational Definitions of Driving Performance Measures and Statistics (Draft), SAE Recommended Practice J2944, 2012.

Copyright © 2013 Victor Ei-Wen Lo and Paul A. Green. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder

Related articles

Article of the Year Award: Outstanding research contributions of 2020, as selected by our Chief Editors. Read the winning articles.