Speech-Interface Prompt Design: Lessons From the Field

Transcription

Speech-Interface Prompt Design: Lessons From the Field
Speech-Interface Prompt Design: Lessons From the Field
Jerome White
Mayuri Duggirala
New York University
Abu Dhabi, UAE
Tata Consultancy Services
Pune, India
[email protected]
ABSTRACT
Designers of IVR systems often shy away from using speech
prompts; preferring, where they can, to use keypad input.
Part of the reason is that speech processing is expensive
and often error prone. This work attempts to address this
problem by offering guidelines for prompt design based on
field experiments. It is shown, specifically, that accuracy can
be influenced by prompt examples, depending on the nature
of the information requested.
Categories and Subject Descriptors
[Human-centred computing]: Empirical studies in interaction design
General Terms
Human Factors
Keywords
Spoken Web, interactive voice response, prompt design
1.
INTRODUCTION
Voice-based systems, and in particular those based on interactive voice response (IVR), have played an influential
role in technology geared for human development. Within
this setting, voice-based systems have been broadly used
for information dissemination [11], community building [8],
real-time monitoring [4], and data collection [9]. They have
been applied, specifically, for purposes as diverse as health
care [10], journalism [7], agriculture [8], and education [6].
In a recent international conference focusing on ICTD [12],
in fact, 20 per cent of accepted work centred around some
application of the voice modality.
Mobile interfaces driven by speech are generally restricted
by inaccuracy: overall task performance is hindered by recognition errors, which in turn negatively affect users’ task
completion and experience. Although this is the nature of
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that copies
bear this notice and the full citation on the first page. To copy otherwise, to
republish, to post on servers or to redistribute to lists, requires prior specific
permission and/or a fee.
ICTD ’15 May 15–18, 2015, Singapore, Singapore
Copyright 2015 ACM 978-1-4503-3163-0/15/05 ...$15.00.
http://dx.doi.org/10.1145/2737856.2737861
[email protected]
speech, quality can be improved by using professional voice
recognition software. The expense of such software, however,
makes the solution unrealistic for small organisations operating on limited resources. For organisations that can afford
such solutions, local languages within developing regions are
often unsupported. Thus, for speech to be a realistic option
within the IVR space, and in turn IVR to be a more viable option within the development community, affordable
improvements to recognition technology are paramount.
One such approach is through prompt design that encourages recognisable input. That is, giving enough dialogue such that a user’s utterance is likely to be within the
recognised vocabulary. Studies around such approaches exist [2, 5, 3], but none have done so on a large-scale, live
deployment, within a developing region. This work fills that
gap. Specifically, it evaluates differences in spoken user input as a function of a prompt’s instructional dialogue. The
results suggest that, depending on the nature of the requested information, a prompt can influence user input, and
in doing so ultimately improve accuracy.
2.
2.1
BACKGROUND
The Employment Service
This work is based on interaction data from a voice-based
employment platform. The employment service allowed candidates to input their resume information, employers to input their job details, and for the two parties to be matched
where appropriate. Candidates were able to apply for jobs
of interest, and employers were able to obtain contact information for candidates they deemed appropriate. The entire
interaction—leaving information, editing information, and
searching for matching parties—took place over the phone.
Users were taken through a series of keypad, known as dualtone multi-frequency (DTMF), and speech prompts to acquire their information and navigate through the system itself. Interaction took place in either English or Kannada,
the local language of the deployment region. Language was
chosen at the start of the call, and interaction took place
in that language for the remainder of the call. The application was built using the Spoken Web platform, a spoken
dialogue system that provides Internet-like features on top
of IVR [1]. Although the service was designed for use on
phones in general, it was intended for use on low-end mobile
phones in particular.
2.2
System Specifics and Definitions
When a candidate called for the first time, they were re-
Table 1: Prompt transcription.
Prompt content
Section
Original
Altered examples
No examples
Location
“Speak the name of the district
where you live. For example, you
may say Mysore, Mandya, Bijapur, Dharwad, et cetera.”
“Speak the name of the district
where you live. For example, you
may say Kolar, Hassan, Gulbarga, Belgaum, et cetera.”
“Speak the name of the district
where you live.”
Skill
“Now, using a single phrase, inform your skills. For example
you may say data entry operator, DTP, plumber, electrician, welder, et cetera.”
“Now, using a single phrase, inform your skills. For example
you may say secretary, waiter,
driver, mechanic, technician,
et cetera.”
“Now, using a single phrase, inform your skills.”
Qualification
“Speak your highest educational
qualification. For example you
may say ITI, diploma, Below
SSLC, PUC, BA, et cetera.”
“Speak your highest educational qualification. For example
you may say B-com, SSLC,
12th standard, B-tech, PGdiploma, et cetera.”
“Speak your highest educational
qualification.”
quired to register with the system. The registration process
consisted of a series of question-answer dialogues that resulted in the creation of a user resume. For purposes of this
paper, an entire question-answer dialogue is referred to as a
section, where the “question” portion of a section is known
as a prompt.
The entire registration process consisted of seven sections:
two sections—a user’s age and work experience—required
DTMF input while the remaining five sections required spoken input. Of the five spoken sections, two, name and free
speech,1 were saved for presentation to job providers—no
speech recognition was attempted in these cases. For the remaining three sections—qualification, skill, and location—
an attempt was made to recognise what the user said. In
an ideal case, the speech was recognised, converted to text,
and stored in a database. For cases in which the speech was
not recognised, a null value was stored, and the user’s input was recorded and saved for offline analysis. A user was
given two attempts to speak an utterance that the system
recognised; the offline recording was made from the third
and final attempt.
2.3
User demographics
The system was deployed throughout the state of Karnataka for a total of 11 months. In that time, candidate
registration was attempted 29,152 times; 8,625 of those attempts completed all sections of the registration process.
Most callers were in their early twenties with one to two
years of work experience. Approximately 21 per cent specified that they were from Bangalore, the largest city in Karnataka; there were over 20 other districts across the state
that were represented by more than 1 per cent of callers,
respectively. The educational split was more uniform, with
most callers possessing at least 10 years of education.
3.
SETUP
As previously mentioned, three sections of the registration process attempted to recognise what the user said and
1
During “free speech,” the user was instructed to speak freely
about themselves for thirty seconds.
then convert that value to text. The prompt for these three
sections instructed the caller not only on the purpose of the
section, but on appropriate words that they should speak.
Specifically, the user was presented with the sections purpose, along with examples of valid input; see Table 1 for details. The basis for the examples was twofold: first, to give
the user an idea of the type of input that was expected. That
is, both to reinforce what was meant by “skill” or “qualification,” for example, and to convey that they should condense
their answer to one or two words. The second reason for
examples was to increase the chances that what a user said
was in the speech recognisers vocabulary. Thus, examples
were provided that were expected to be recognised.
The examples chosen for the original prompts were done so
based on field studies [13] and discussions with government
employment specialists; thus, they were chosen for their relatability to the target audience and designed in conjunction
with the speech recognition dictionaries. The examples for
the experimental prompts were chosen based on what was
already in the recogniser vocabulary. They were chosen at
random, with the only condition being that the system was
capable of recognising them if spoken and that they were not
already present in the original prompt. Equivalent prompts
were run for both Kannada and English.
The original prompts ran in the system, exclusively, for 10
weeks. From this, a reference distribution of user input was
established. After this period, users received one of the three
prompts—original, altered examples, or no examples—from
a random uniform distribution. Original prompt statistics
are reported from the start of the experimental time period,
not from system inception.
4.
RESULTS
Results from the experimental study are outlined in Table 2. Three aspects of the data are of particular interest:
the amount of time required to go through each section,
the accuracy of speech recognition, and the distribution of
recognised answers. Reported times are exclusively user interaction time—the amount of time spent listening to the
prompt itself is removed.
Table 2: Prompt effects on user input. N is the number of samples taken for a given prompt. Time
refers to the amount of time users spent in a given prompt; stars (?) denote values that are signficantly
different from “original” (two-sided paired t-test, p < 0.05). Accuracy refers to the amount of user input that
was successfully recognized. Ex. Dist. (“example distribution”) is cumulative distribution of the examples
mentioned in the prompts. Top-five inputs are the most common system-recognized input given by users,
presented in descending order; daggars (†) denote input values were also used as examples in the prompt.
Prompt Statistics
Time (sec.)
Section
Prompt
N
Avg
SD
Accuracy
Ex. Dist.
Location
Original
1597
33.3
22.1
73.6%
0.179
Altered Examples
1550
32.4?
19.8
75.8%
0.112
No Examples
1506
29.9?
24.0
69.7%
Original
1915
51.7
29.6
38.3%
0.155
Altered Examples
1953
45.1
33.6
50.5%
0.281
No Examples
1938
51.0?
30.9
18.7%
Data entry operator† , Electrician† ,
DTP† , Attendant, Fitter
Technician† ,
Secretary† ,
Teacher,
†
Driver , Mechanic†
Computer Science, Electrician, Cashier,
Finance
Original
Altered Examples
No Examples
2140
2101
2174
43.8
45.9?
38.0?
24.7
26.7
26.3
64.2%
62.3%
58.8%
0.331
0.128
PUC† , BA† , ITI† , BEd, BCom†
BA, ITI, BCom† , BEd, 12th standard†
PUC, BA, ITI, BCom, BEd
Skill
Qualification
4.1
Input bias
Example distribution (Table 2, “Ex. Dist.”), and top-5 input, are telling indicators of whether users are merely repeating the examples they hear. Example distribution is the
fraction of users exposed to a given prompt who responded
with a value that was also present in the example. Based
on values observed across all prompts, it is unlikely that the
results are dominated by repeats. If this were the case, observed values would have been closer to one—in no section,
in fact, was the cumulative distribution of the prompt examples greater than 35 per cent. Further, in no prompt did the
top-5 values observed coincide completely with the values
presented in the prompt examples. These findings suggest
that the values extracted during this experiment were actual
representations of the population.
4.2
Accuracy
In all cases, irrespective of section, the accuracy of the
speech recogniser was lowest when no examples were presented to the user. The gain in accuracy, however, was not
constant across sections: whether a gain was observed if
examples were presented depended on the section, and ultimately the nature of the information extracted from the
user. In the case of location and qualification, an increase of
4 and 2.8 percentage points, respectively, were observed, in
the worst case.2 In the case of skill, however, accuracy doubled with the introduction of examples. Yet, the best case
accuracy of any skill example section was almost 10 percentage points lower than the no-example accuracies of location
2
Where “worst case” is the difference between the noexample accuracy and the lower of the original and altered
example accuracies.
Top-Five Inputs
Bangalore, Bijapur† , Bellary, Dharwad† ,
Bagalkot
Bangalore, Bijapur, Bellary, Belgaum† ,
Dharwad
Bangalore, Bijapur, Bellary, Bagalkot,
Hassan
and qualification; which happened to be the worst case observed accuracies in those sections.
It is likely that skill accuracy was poor, and that qualification was slightly worse than location, because of the nature of the information. When asked to describe their skill,
many people will either list several concise topics, or speak
freely, in a conversation-like manner, about a single ability. Moreover, what constitutes a skill, and how to describe
it, are often unique across a populous: the same expertise
may be described several different ways depending on who
is speaking. This combines to make skill a difficult concept
to classify, and in turn to recognise automatically.
Within the location section, regardless of whether examples were presented, the user was asked to speak the name
of their “district,” an administrative division in India, that
is well-defined and widely understood. There are approximately 700 districts within the country, 30 of which are
contained in Karnataka. This not only simplifies the responsibility of the speech recognition engine, it reduces the
values considered by the caller to a set that is finite and universal. While there may be several ways to describe skill,
and slight variations in degrees offered across syllabi, districts are relatively well known and stable.
Qualification accuracy, overall, was better than the accuracies observed for skill, but worse than those observed for
location. Qualification benefits from many of the same advantages that were seen in location: a finite, well-established
set of values, which was further reduced given that the platform had a target audience. However, qualification—like
skill, but to a lesser extent—is plagued by a lack of standardised structure, along with various methods of expression.
India has various boards that govern primary, secondary,
and trade-specific tertiary education. As such, equivalent
degrees may go by different names depending on the board
governing a particular student’s education. Complicating
matters further, some people may express their last educational instance as a combination of that instance, and the
instance result: as opposed to “PUC,” for example, a person
may use what is a more accurate description of “PUC-fail.”
They may also combine their last successful degree and the
degree that they are currently pursuing, making the speech
interaction more conversational and harder to automatically
discern.
4.3
Time
With the exception of skill, the amount of time required to
move through a section was smallest when no examples were
presented. The difference is slight—about five seconds in the
worst case—but observable. This increase in section time,
however, generally led to accuracy improvements. Skill was
an exception: the duration of the no-example prompt was
longer, and the accuracy was much lower, when compared
to the original and altered example prompts.
5.
CONCLUSION
What examples to use within a prompt, and even whether
to use examples at all, depends on the nature of the data
and the acceptable trade-off between time and accuracy. In
all cases, there was an improvement in accuracy when examples were introduced. There was also—with the exception
of skill—an increase in time spent to move through the section. However, whether the increased time was worth the
improved accuracy was not always clear. This is most notable in the case of location: examples, in the best case,
improved accuracy by over 6 percentage points. Inclusion
of these examples, however, cost the caller, and the system,
an additional 2.2 seconds of connection time, on average.
Moreover, irrespective of which examples were used, or if
examples were used at all, there was little difference between
the observed top-five inputs. Again, this suggests that location is relatively standard in the minds of callers, and that
creating prompts that optimise time would not significantly
hinder the quality of the data collection.
Qualification benefits from examples, but the time required for those examples could likely be reduced; presenting
fewer examples than were presented in our system, for example, would be one way of achieving this. Again, from the
overlap in top-five inputs across example schemas, what a
qualification is seems clear to our user group; examples likely
serve to educate them on what format—standard one-word
abbreviations—is expected.
The overwhelming conclusion for skill is that examples
are not just helpful, but necessary. As noted in Table 2,
the discrepancy between the best case and worst case skill
prompts—no examples and altered examples, respectively—
was the largest across all prompts and sections in which
the experiment was conducted. Further, that users spent as
much or more time in the no-example case as they did in the
original and altered example cases, suggests that either their
speech was too long, or that it required numerous retries to
move through the prompt.3 Further, what was particularly
interesting was that in the case where examples were cho3
And even then, “success,” in the sense of the system recognising their input, was not guaranteed.
sen at random (“altered examples”) the accuracy was higher
than in the case where the examples were chosen based on an
understanding of the potential user base (“original”). This
suggests that for relatively open-ended data collection—as
is the nature of skill—systems, and researchers, should constantly adapt and observe the prompt to maximise success.
6.
ACKNOWLEDGEMENTS
The authors would like to thank the Karnataka Vocational
Training and Skill Development Corporation for their assistance in making this work possible; Vaijayanthi Desai and
Kundan Shrivastava for their infrastructural assistance; and
Brian DeRenzi whose meaningful discussions played a key
role in the formation of this work. Appreciation is also extended to the reviewers for their insightful comments and
suggestions.
References
[1] S. Agarwal et al. The spoken web: a web for the underprivileged. SIGWEB Newsletter, (Summer):1:1–1:9,
June 2010.
[2] C. Baber et al. Factors affecting users’ choice of words
in speech-based interaction with public technology. Int.
J. Speech Tech., 2(1):45–59, 1997.
[3] K. Baker et al. Constraining user response via multimodal dialog interface. Int. J. Speech Tech., 7(4):251–
258, 2004.
[4] W. Curioso et al. Design and implementation of CellPREVEN: a real-time surveillance system for adverse
events using cell phones in Peru. In AMIA Annual Symposium, 2005.
[5] L. Karsenty. Shifting the design philosophy of spoken
natural language dialogue: From invisible to transparent systems. Int. J. Speech Tech., 5(2):147–157, 2002.
[6] M. Larson et al. I want to be Sachin Tendulkar!: A spoken English cricket game for rural students. In CSCW,
2013.
[7] P. Mudliar et al. Emergent practices around CGNet
Swara, voice forum for citizen journalism in rural India.
In ICTD, 2012.
[8] N. Patel et al. Avaaj Otalo: a field study of an interactive voice forum for small farmers in rural India. In
CHI, 2010.
[9] S. Patnaik et al. Evaluating the accuracy of data collection on mobile phones: A study of forms, SMS, and
voice. In ICTD, 2009.
[10] J. Sherwani et al. Healthline: Speech-based access to
health information by low-literate users. In ICTD, 2007.
[11] J. Sherwani et al. Speech vs. touch-tone: Telephony
interfaces for information access by low literate users.
In ICTD, 2009.
[12] B. Thies and A. Nanavati, editors. Proceedings of the
3rd ACM DEV, 2013.
[13] J. White et al. Designing a voice-based employment
exchange for rural India. In ICTD, 2012.