Listening In - MIT Center for Digital Business

Transcription

Listening In - MIT Center for Digital Business
A research and education initiative at the MIT
Sloan School of Management
“Listening In” to Find Unmet Customer Needs
and Solutions
Paper 156
Glen L. Urban
John R. Hauser
January 2003
For more information,
please visit our website at http://ebusiness.mit.edu
or contact the Center directly at [email protected]
or 617-253-7054
“Listening In” to Find Unmet Customer Needs and Solutions
Glen L. Urban
and
John R. Hauser
January 3, 2003
Glen L. Urban is the David Austin Professor of Marketing, Sloan School of Management, Massachusetts Institute of Technology, E56-332, 38 Memorial Drive, Cambridge, MA 02142, (617)
253-6615, fax (617) 253-7597, [email protected].
John R. Hauser is the Kirin Professor of Marketing, Sloan School of Management, Massachusetts Institute of Technology, E56-314, 38 Memorial Drive, Cambridge, MA 02142, (617) 2532929, fax (617) 253-7597, [email protected].
This research was supported by the Sloan School of Management, the eBusiness Center, the Center for Innovation in Product Development at M.I.T., and the General Motors Corporation. We
gratefully acknowledge the contributions of our industrial collaborators, research assistants, and
faculty colleagues: Vince Barabba, Iakov Bart, Ahmed Benabadji, Rupa Bhagwat, Brian Bower,
Brian Chan, Hann-Ching Chao, Mitul Chatterjee, Shyn-Ren Chen, Thomas Cheng, Stanley
Cheung, Frank Days, Ken Dabrowski, Benson Fu, Salman Khan, Christopher Mann, Rami Musa,
Joseph Kim, Ken Lynch, Bill Qualls, James Ryan, Bilal Shirazi, Jonathon Shoemaker, Fareena
Sultan, Andy Tian, Xingheng Wang, and Irene Wilson. This paper has benefited from presentations at the Marketing Science Conferences in Wiesbaden Germany and in Edmonton, the MIT
Marketing Workshop, the New England Marketing Conference, and the Stanford Marketing
Workshop. An Adobe Acrobat version of this paper is available at mitsloan.mit.edu/vc.
“Listening In” to Find Unmet Customer Needs and Solutions
Abstract
In many industries, such as the automotive, travel, and health industries, customers routinely use the Internet to gather information on product features and to refine their choice sets.
As a result, many websites (e.g., Kelley’s Blue Book) now provide virtual advisors to help customers narrow their searches. Customer preference data from these virtual advisors are available
at little incremental cost and provide a natural source of ideas for new product platforms.
In this paper, we explore a practical marketing research methodology to identify new
high-potential unmet-need segments for product development by “listening in” to these ongoing
dialogues between customers and web-based virtual advisors. The methodology is designed to
be practical when there are large numbers of options (e.g., as many as 150 trucks when studying
the automotive market) and many potential customer needs. Calibration is based on activities,
interests, and opinions data (AIO) that are collected routinely by many firms. For example,
automotive firms collect AIO data periodically for a variety of purposes. Calibration can evolve
with incremental data collection as new products enter the market and as new customer needs are
identified. The methodology, designed to complement extant methods, uses five modules – a
Bayesian virtual advisor, a listening-in trigger, a virtual engineer, a design palette, and a mechanism to identify the underlying unmet-need segments. We also provide an “opportunity-sizing”
method which takes product costs into account and provides a rough estimate of the size of the
opportunity using available data. The modules are flexible; most can be used with other extant
methods.
We describe the methodology and examine its properties through formal analysis and
Monte Carlo simulation. We demonstrate feasibility by applying the methodology to the pickuptruck category (over 1,000 web-based respondents). The methodology identified two ideas for
new truck platforms worth approximately $2.4-3.2 billion and $1-2 billion, respectively.
“Listening In” to find Unmet Customer Needs and Solutions
Identifying new platform opportunities is one of the most important roles of market intelligence.
Monitoring [web-based advisors] provides a rich source of observed in-market customer behavior that
complements our current inquiry tools which, by their nature, are forced to ask customers either to state
their intentions before they are actually in the market or to remember after the purchase what they did
(and why) when shopping for a vehicle. No form of inquiry is perfect, however, whatever its limitations,
the currency [of web-based advisors] presents a valuable source of market understanding that is already
streaming by and is of great value when used appropriately.
–
Vince Barabba, General Motors
General Manager of Corporate Strategy and Knowledge Development
(responsible for overseeing GM’s New Business Development Network)
Unmet Customer Needs Represent New Opportunities
Identifying unmet-need opportunities is an important marketing problem in a variety of
industries. For example, in the US truck market, a single share point is worth $800 million in
annual revenues. One of the top priorities of strategic marketing is to identify promising new
segments of the market with unmet needs that can be fulfilled with a new truck platform. If such
opportunities are confirmed, these segments justify investments of $1-to-2 billion in product design and engineering. Even a one- or two-percent segment is worth further investigation if the
needs of the segment can be fulfilled with a minor redesign (typical cost of $300 million).
With so much at stake, automotive manufacturers have been innovators in marketing research and market evaluation methods. General Motors (GM) alone spends tens of millions of
dollars annually to understand the voice of the customer and to design products to meet customer
needs (private communication – exact numbers confidential). Methods such as focus groups,
ethnographic methods, means-ends chains, “clinics,” prelaunch forecasting methods, information
acceleration, and conjoint analysis are used frequently. One GM supplier alone reported to us
that they used conjoint analysis methods to gather data on feature levels from over 10,000 customers in the previous year. Automotive manufacturers invest heavily in such studies when new
segment(s) are identified, especially if potential product features relevant to that segment are
known. These marketing research methods are based on a wealth of experience. For example,
conjoint-analysis-like studies have been used routinely by General Motors since the 1930s (Time
1938). For other examples see Crawford 1991; Dolan 1993; Green and Srinivasan 1990; Griffin
and Hauser 1993; Gutman 1982; Kolpasky 2002; Lehmann and Winer 1994; Narasimhan and
Sen 1983; Shocker and Srinivasan 1979; Urban and Hauser 1993; Urban, Hauser, and Roberts
1
“Listening In” to find Unmet Customer Needs and Solutions
1990; Urban, Weinberg, and Hauser 1996; Wind 1982; and Zaltman 1997. To complement experience in researching identified opportunities, senior automotive managers seek improved
methods to identify new opportunities.
In this paper, we explore new marketing research methods to uncover, understand, and
evaluate promising opportunities for new automotive products. In the parlance of the industry,
we seek new “fishing grounds” based on unmet customer needs. This managerial problem is not
new, nor is it limited to the automotive industry. Product development teams have faced this
marketing problem for many years and continually explore new methods (Barabba 2003;
Barabba and Zaltman 1991). We do not seek to replace extant market research methods, but
rather complement them by analyzing data that are collected for other goals. Specifically, we
exploit new data that are obtained by “listening in” to ongoing “dialogues” created when customers use the Internet to search for information and advice about automotive purchases.
These data are attractive for identifying new needs-based segments. First, the data
collection is incentive compatible – customers are seeking advice and, thus, have an incentive to
reveal their needs. Second, the data are available at little incremental cost – we need only monitor these dialogues electronically as they occur. Third, virtual advisors are updated often to include new vehicles and new customer needs, providing evolving data with which to identify new
unmet needs as soon as customers express those needs. Fourth, data from virtual advisors have
the scale necessary for opportunity identification in the automotive industry. There are approximately 150 pickup truck sub-brands on the market; each major manufacturer offers over 20,000
brand-feature combinations. (A typical segmentation study might analyze questionnaire responses from 100,000 customers and identify almost fifty segments.) Virtual-advisor data match
this scale. For example, one virtual advisor, sponsored by GM, J. D. Power, and Kelley’s Blue
Book, has approximately 350,000 annual visitors.
However, we must overcome challenges if the methodology is to be practical. Virtual
advisors are based on needs and features that are known – we must develop a method that looks
into the gaps. With the large number of vehicles, segments, and customer needs, the method
must scale well to handle large problems and evolve as new needs and new vehicles are added to
the virtual advisors. We must be able to estimate underlying parameters (calibrate the model)
based on data that are now collected routinely by automotive manufacturers. If this alone were
not a challenge, we must gather this information without compromising the primary function of
2
“Listening In” to find Unmet Customer Needs and Solutions
the virtual advisors – the method must not be overly intrusive. Questions should be directed,
brief, and seen as relevant by the customers. Finally, because of the large scale and continuous
nature of the data, the method should be automated. The method we propose attempts to address
each of these implementation issues.
We begin by defining the marketing research problem. We then describe a representative
virtual advisor that uses Bayesian methods to focus quickly on key customer needs. (Listeningin methods can also work with other advisory systems.) Next we identify a “listening-in” trigger
mechanism which senses unmet needs and invites a virtual engineer to join the dialogue with
customers whose needs are not met with existing vehicles. The virtual engineer engages the customer by asking directed questions to explore the potential unmet need. The customer also expresses his or her needs by using a design palette to create virtual products that are not now on
the market. These automated methods gather data from each customer as that customer visits the
virtual advisor. Periodically, the strategic marketing group reviews the data to summarize opportunities. Cluster analysis provides a rough estimate of the size of any identified opportunity – an
estimate that will be clarified later with seed investments and extant marketing methods.
After introducing the listening-in method we use Monte Carlo simulation to demonstrate
the method, explore sensitivity to various trigger levels, and explore sensitivity to response errors. We then illustrate the listening-in methods with an application based on over 1,000 respondents. The application identified promising new truck concepts for a major automotive manufacturer. Although we focus on the automotive industry, we expect that other industries share many
of the same characteristics that make “listening in” valuable. For example, the fraction of customers who seek information and advice from the Internet is 70% in travel services and 56% in
health services. Both industries have the scale necessary to make “listening in” feasible.
Marketing’s Role in Finding New Opportunities for Product Development
Marketing and marketing research interact with product-development teams at many
stages in the phase-review, stage-gate, or waterfall processes of opportunity identification, concept creation, design and engineering, testing, and launch. For example, voice-of-the-customer
methods, focus groups, and ethnographic approaches all explore opportunities once they are
identified and provide a lens on the needs of a customer segment (e.g., Burchill 1992). During
the “design and engineering” phase conjoint analysis clarifies and prioritizes feature levels. During the “testing” phase of product development, information acceleration methods test virtual
3
“Listening In” to find Unmet Customer Needs and Solutions
concepts and clinics test new physical prototypes. Once needs-based segments are defined, the
process is effective and efficient. But marketing’s role is not limited to existing market segments.
As consumer tastes evolve and new technologies make new features feasible, promising
unmet-need segments emerge. If the segment is large and the needs are truly unmet, such opportunities can revolutionize the industry. These opportunities are extremely profitable for the innovating firm. For example, in the mid-1960s Ford identified the trend of teenagers and young
adults to customize inexpensive vintage Fords with V8 engines. To meet this opportunity they
launched the 1964½ Mustang, which captured the hearts of a new generation of baby boomers
just reaching driving age (Figure 1a). This small, inexpensive sports car with a powerful V8 engine sold 420,000 units in the first year ($10 billion at today’s prices) and went on to launch the
lucrative “pony” segment (classicponycars.com/history.html). In 1983 Chrysler introduced the
Dodge Caravan and Plymouth Voyager minivans – downsized vans for now-growing families,
built on a car-like K platform with comfort features such as power windows, locks, seats, and
quality sound systems (Figure 1b). This new vehicle could carry a 4’x 8’ sheet of plywood but,
unlike existing vans, could fit easily in customers’ garages, drove like a passenger car, had a side
door for small children, and had a sedan-like liftgate for shopping. Minivans fulfilled these
needs by exploiting front-wheel drive to avoid high floors and to avoid an engine that tunneled
into the cabin. Chrysler sold 210,000 units in the first year and dominated this new segment for
years to come (allpar.com/model/m/ history.html).
Figure 1
Examples of Significant New Automotive Platform Opportunities
(a) 1964½ Ford Mustang
(b) 1983 Dodge Caravan
With so much as stake, the strategic marketing and marketing research groups invest
heavily in identifying new opportunities. They speak to leading edge users, maintain and monitor user groups, sponsor special racing events, monitor chat rooms and user groups, and use a
4
“Listening In” to find Unmet Customer Needs and Solutions
variety of qualitative methods (Barabba 2003). However, given the scale and scope of the automotive industry, marketing specialists at automotive manufacturers continually explore new data
sources and new methods to identify potential unmet-need segments.
The marketing task is often integrated with a design-and-engineering team. With the
marketing data in hand, the team develops a “platform” to exploit economies of scale in both engineering and manufacturing. The platform may span several brand names providing a balance
between differentiated “positionings” and a common technological base. A typical automotive
platform, say a light-pickup-truck platform, requires well over one billion dollars in investment
for new capital (tools and assembly), expendable materials (hand-built prototypes), and engineering time. Because a typical platform redesign requires as many as 1,200 person-years of engineering effort, marketing bears substantial responsibility for assuring that the new platform opportunity has sufficient revenue potential.1
With such responsibility, marketing professionals at automotive manufacturers undertake
many parallel analyses to find (and confirm) new opportunities. Such methods reinforce one another. For example, ethnographic methods might identify a set of features to be added to the virtual advisor, the analysis of which, in turn, identifies a new opportunity. The opportunity is then
clarified with still other analyses.
Tapping Data from Virtual Advisors (Web-based Searches)
The data obtained by monitoring web-based customer searches are extensive and provide
the scale necessary to explore complex categories such as trucks. However, such data are relatively underutilized as a source of unmet-need opportunities. Websites such as Kelley Blue
Book (kbb.com), Microsoft Autos (autos.msn.com), Edmund’s (edmunds.com), Auto-by-tel
(autobytel .com), Autoweb (autoweb.com), NADA (nadaguides.com), and Vehix (vehix.com)
have changed the way that customers search for information on cars and trucks. Sixty-two percent (62%) of all new-vehicle buyers search on-line before buying a vehicle (J. D. Power 2001).
This search rate has increased from 54% in 2000 and 40% in 1999. According to J. D. Power,
“the average automotive Internet user visits 6.8 automotive sites before purchase – 80% visit at
least one independent site and 71% visit at least one manufacturer Web site.” The mostimportant and most-accessed Internet content was information about vehicle options and features. Interestingly, while customers prefer independent sites for pricing and general evaluation,
5
“Listening In” to find Unmet Customer Needs and Solutions
they prefer manufacturers’ sites, by more than a two-to-one margin, for feature and option information (J. D. Power 2001, p. E16). Such searches benefit consumers. Scott Morton, Zettelmeyer, and Silva-Risso (2001) report that Auto-by-tel customers pay approximately $450 less for
their car when controlling for dealer choice and trade-ins.
While the data from web-based searches are a natural source of information, the information flow is extensive. To use these data we need a system to filter systematically the information about unmet needs that is embedded in the data. Thus, we examine first those web-based
searches that are most likely to contain embedded data on unmet needs – searches directed by a
virtual advisor. Virtual advisors seek to provide unbiased information on vehicle features and
options and, in doing so, help customers find the vehicle or option that is right for them (Urban,
Sultan and Qualls 2000). Some virtual advisors are independent (e.g., AOL’s New Car Guide at
aolsvc.carguides.aol.com/cars/car_guide.jsp or ActiveBuyersGuide.com) while others are sponsored by manufacturers (e.g., General Motors’ autochoiceadvisor.com and Michelin-UniroyalBFGoodrich’s tireadvisor.com).
If the website is successful in building trust, then customers articulate needs accurately.
Usually, the virtual advisor can recommend a product to fulfill a customer’s needs, but not always. When the advisor cannot recommend an existing product a promising opportunity might
exist. In this paper we derive a trigger to identify automatically when virtual advisors cannot
recommend existing products because such products do not fulfill consumers’ needs.
Virtual advisors come in many varieties and are used on the web in categories such as
travel, health, autos, computers, home entertainment, and financial services. One approach is
represented by “comparators” that array choice alternatives by features (ePinions.com). Another
approach is feature-specification which asks the consumer for preferred levels of features and
searches the data base for products that meet the feature specifications (Kelly Blue Books’ recommendation tool – kbb.com). Configurators develop detailed specifications and costs for the
chosen set of detailed product features (configurator.carprices.com/autoadvisors/). Other systems
use collaborative filtering to recommend products based on correlations of past and future purchases for similar customers (amazon.com) and some advisors use real people accessed by email (mayohealth.org). Others use live chat rooms to exchange information (nordstom.com).
Virtual advisors are evolving, especially in the automotive industry. GMBuypower.com uses an
1
Private communications on issues and data with automotive executives at various manufacturers.
6
“Listening In” to find Unmet Customer Needs and Solutions
algorithm to weigh features and preferences to identify a maximum-utility alternative. ActiveBuyersGuide.com uses an underlying conjoint-analysis-like procedure to determine preferences
and calculate utility values for real products based on questions about tradeoffs between profiles
of features. Other advisors mix questions about features and characteristics with questions about
how the vehicle will be used (the Bayesian advisor described below).
Adaptations of the listening-in methodology should work with many of these (trusted)
virtual advisors, but not all. To work with the listening-in methodology, an advisor must have
two key characteristics. The virtual advisor should be adaptive in the sense of being capable of
providing recommendations to the customer at any point in the questioning sequence.2 In addition, it must be feasible to link the customer’s responses to descriptions of vehicles – although
this linkage can be indirect. We illustrate the listening-in methodology for a Bayesian virtual
advisor, but the basic ideas are readily adaptable to feature-specification advisors and utilitymaximization advisors, including conjoint-analysis-based advisors. In theory, we could monitor
the click-streams in configurators to identify unexpected feature combinations, but, at present,
configurators do not have any companion predictive systems for a Bayesian trigger. Because the
human interface would need to be addressed for the real-people and chat-room advisors – such
advisors might be best monitored by traditional, rather than automated, means. The listening-in
methodology is unlikely to be applicable to collaborative filtering until a method can be developed to link filtering to product characteristics.
A Bayesian Virtual Advisor
We now illustrate the listening-in methodology with a virtual advisor for truck purchases
that was developed as a prototype for a major automotive manufacturer. (A commercial system
based, in part, on key concepts from this advisor is now in place on the web.) The virtual advisor
combines two methods to recommend a set of four vehicles to customers – a segmentation
“gearbox” and a Bayesian advisor. The segmentation gearbox divides people into segments
based on grouping and assignment rules.3 In this case the grouping is based on a cluster analysis
of a 114-item AIO (Activities, Interests, and Opinions) questionnaire sent to 100,000 respondents (76 personal viewpoints and 38 preferred vehicle characteristics – including styling and
2
Many advisors require a baseline set of questions before entering an adaptive phase. Most virtual advisors are
adaptive because web-based customers have proven to be impatient with lengthy questions on features or needs.
3
The colorful industry term, gearbox, is an analogy. Just as the gearbox in a car matches engine speed to wheel
speed, the segmentation questions match the manufacturer’s vehicles to the customer.
7
“Listening In” to find Unmet Customer Needs and Solutions
design). See Plummer (1974) for a general description of AIO questionnaires. The original
study by the automotive manufacturer identified forty-eight segments of which twenty-five were
relevant to pickup trucks. Customers were assigned to segments based on answers to the virtual
advisor’s questions – answers about the customer’s desire for features and options such as comfort, passenger capacity, and prestige as well as about the customer’s anticipated use of the truck.
The answers to these questions assign respondents to a group of similar people. In the virtual
advisor one of the four recommended vehicles is the vehicle bought most often by the segment to
which the customer is assigned. Like collaborative filtering, the segmentation gearbox provides
reasonable recommendations based on preferences of similar customers. It is based on extant
marketing research methods, but it does not identify new opportunities.4 Instead, we focus on
the Bayesian advisor which is used to recommend three of the four vehicles.
Bayesian Advisor
The basic concept behind the Bayesian advisor is (1) to select sets of questions, known as
question banks, such that the answers provide the most information about which vehicle to recommend and (2) after each question bank to update the probabilities that describe the likelihoods
that each vehicle will be most preferred by the customer.5 Figure 2a illustrates the opening
screen of the virtual advisor (a neighbor who has bought many trucks over the years) and Figure
2b illustrates one of the question banks asked of customers. We describe first the Bayesian updating mechanism and then describe how this mechanism can be used to select the maximuminformation question bank. We later indicate how the conditional probabilities and prior probabilities are obtained.
The Bayesian advisor shares some characteristics with tailored interviewing (Balasubramanian and Kamakura 1989 [BK]; Singh, Howell, and Rhoads 1990 [SHR]; Kamakura and
Wedel 1995 [KW]). As in the Bayesian advisor, questions in tailored interviewing are selected
to maximize an information measure and thus reduce the number of questions that need be asked.
4
Gearboxes can be made more efficient with tailored interviewing techniques (Balasubramanian and Kamakura
1989; Singh, Howell, and Rhoades 1990; Kamakura and Wedel 1995). For example, Kamakura and Wedel demonstrate that a tailored procedure based on latent-class models can classify customers to segments with two-thirds the
number of questions that were necessary with a clustering-based gearbox. Because the gearbox is not the focus of
this paper we leave such improvements to future research.
5
The global set of question banks, from which the algorithm selects, is drawn from cluster analyses of the ongoing
AIO surveys, supplemented with managerial judgment. The set of question banks evolves based on ongoing market
intelligence. The methods to identify the set of question banks are state-of-the-art, but standard, marketing research
practice. They are not the focus of this paper.
8
“Listening In” to find Unmet Customer Needs and Solutions
In both methods data for key parameters are obtained from prior questionnaires. However, the
goals and scope differ. Tailored interviewing seeks to select questions from a larger bank either
to uncover an unidimensional latent trait (from dichotomous scales in Balasubramanian and Kamakura 1989 or Likert-type data in Singh, Howell and Rhoads 1990) or to assign customers to
segments based on the bank of questions (seven segments in Kamakura and Wedel 1995). The
Bayesian advisor focuses on recommending vehicles and, because of the scale in the truck market (148 vehicles in our application), the advisor must bootstrap these recommendations with
Bayesian updating. Nonetheless, it is theoretically possible to develop an advisor based on tailored interviewing and such an advisor would be compatible with the listening-in methods.
We begin with the notation. Let Q be a set of question banks indexed from q = 1 to N.
For each question bank, q, let rq index the potential responses to that question bank where rq is a
nominal variable with values from 1 to nq. If there is more than one question in a question bank,
then nq represents the number of possible combinations of answers. If one of the questions includes a continuous sliding scale, it is discretized to a finite number of categories.
Figure 2
Example Question Banks Asked by a Virtual Advisor
(a) Introductory Screen
(b) Example Question bank
For each customer the order of the question banks is chosen adaptively. For a given customer, let Rq-1 be the set of question banks up to, but not including, question bank q. Let vj indicate vehicles from 1 to V. We are interested, at any point in the advisor’s questioning sequence,
in the likelihood that the customer will prefer vehicle j after having been asked question bank q.
We indicate this likelihood by P(vj | Rq-1, rq).
9
“Listening In” to find Unmet Customer Needs and Solutions
Suppose that we have available from earlier surveys, the conditional probabilities of how
customers, who prefer each vehicle, will answer the question banks. Then, we can use Bayes
Theorem to update recommendations.6
(1)
P( v j | Rq−1 , rq ) =
P( rq | v j , Rq−1 ) P( v j | Rq−1 )
V
∑ P( r | v , R
q
j
q −1
) P( v j | Rq−1 )
j =1
where P(vj | Rq-1) was the virtual advisor’s recommendation probability to the customer for vehicle vj prior to asking the qth question bank.
However, even with data from full-scale surveys such as an AIO questionnaire with
100,000 responses, using Equation 1 is not feasible because the number of potential combinations of responses grows exponentially with the number of question banks. For example, in our
study the dimensionality of RN, the number of unique paths through the advisor’s questions, is
1.4 x 1015. Fortunately we can make Equation 1 feasible based on the property of “local independence.” This property appears reasonable for our data and has proven robust in simulations
and applications in the tailored-interviewing literature (e.g., SHR Equation 8; KW Equation 11).
As SHR explain, local independence recognizes that there will be non-zero correlations across
vehicles in the answers to the question banks – those customers who prefer a full-sized truck may
also be likely to prefer a diesel engine. Indeed, it is this combination of preferences upon which
the advisor bases its recommendations. However, if we limit ourselves to customers who prefer
a Ford F350 Supercab, then, for those customers, responses to the “size” question bank are approximately statistically independent of the responses to the “engine type” question bank. This
enables us to write P(rq, Rq-1| vj) ≅ P(rq | vj) P(rq-1 | vj)… P(r1 | vj) which implies that P(rq | vj) ≅
P(rq | vj, Rq) by the laws of conditional probability. Using this property, we rewrite Equation 1 as
follows where P(vj | Rq-1) is obtained recursively:
(1’)
P( v j | Rq , rq ) ≅
P( rq | v j ) P( v j | Rq )
V
∑ P( r | v ) P(v
q
j
j
| Rq )
j =1
Figure 3 gives a simplified example for one customer of the evolution of the recommendation probability. The current recommendation is given on the left and the probability that the
customer will purchase that recommended vehicle is given on the right. Also listed on the left
6
In all equations, we suppress the individual customer subscript, i, for simplicity
10
“Listening In” to find Unmet Customer Needs and Solutions
are the question bank and the answer. For example, after the second question bank on engine
size, the customer answers “4 cylinders.” If the customer were to stop answering question banks
and request a recommendation, the advisor would recommend the Mazda B2300 and forecast a
0.0735 probability that the customer would purchase the Mazda B2300. In Figure 3 the probability of purchase increases for the most preferred truck after each question bank is answered.
Note that the recommended vehicle changes after the fifth question bank and again after the
eighth question bank.
Figure 3
Evolution of Updated Recommendation Probabilities After Question Banks
Recommendation/Question Banks
Mazda B2300, Prior (Points *)
0.0533
Mazda B2300, Engine Size (4 cyl)
0.0735
Mazda B2300, Transmission (Auto, 2WD)
0.0861
Mazda B2300, Size (Compact)
0.1105
Mazda B2300, Towing/Hauling (no)
0.1123
Toyota Tacoma, Construction Plowing (no)
0.1200
Toyota Tacoma, Brand (All)
0.1243
Toyota Tacoma, Bed Length (Short)
0.1328
GMC Sierra 1500, Tallest Person (6'-6.5')
0.1376
GMC Sierra 1500, Passengers (2)
0.1440
GMC Sierra 1500, Maneuverability (Important)
0.1440
GMC Sierra 1500, Big, Quiet (Not Important)
0.1458
GMC Sierra 1500, Styling (Sporty)
0.1467
GMC Sierra 1500, Price (20-22K)
0.1467
0.00
0.04
0.08
0.12
0.16
Maximum Probability
Question Bank Selection
To select the next question bank the virtual advisor attempts to gain as much information
as possible from the customer. For example, if, after reviewing the responses, the advisor decides that a question bank on towing capacity is likely to make one truck more highly probable
and all other trucks less probable, then that question bank might be a good candidate to ask next.
To do this formally, we turn to formal theory in which information is defined as the logarithm of
the relative odds (e.g., Gallagher 1968). That is, the information, I(vj | rq, Rq-1), provided by the
response to question bank q, equals log [P(vj | Rq-1, rq)/ P(vj | Rq-1)]. This definition has a number
of nice theoretical properties including that (1) under an equal proportional loss rule, information
always increases when the probability of the maximum-choice truck increases, (2) the expected
11
“Listening In” to find Unmet Customer Needs and Solutions
information is maximized for the true probabilities, and (3) the information measure rewards systems that provide more finely-grained estimates (Kullback 1954; Savage 1971).
In order to compute the expected information, we need to take the expectation over all
possible responses to question bank q and over all possible vehicles. Thus, the information that
we expect from question bank q is given by Equation 2.7
(2)
V
nq
j =1
rq =1
EI ( q | Rq−1 ) = ∑ P( v j | Rq−1 )∑ P( rq | v j , Rq−1 ) log
P( v j | rq , Rq−1 )
P( v j | Rq−1 )
In a myopic world, the virtual advisor would simply choose the question bank for which Equation 2 is maximized.
We can improve upon Equation 2 with an m-step look ahead. To date, the computational
demands of Equations 1 and 2, coupled with the large number of responses for each question
bank has limited the look-ahead algorithm to two steps. Basically, for each potential question
bank and response on Step 1, the advisor computes the best second question bank and the expected information for that question bank. It then selects the Step-1 question bank with the highest contingent expected information.
Initial Calibration
Two estimates are necessary for the virtual advisor to begin: prior probabilities, P(vj), and
the conditional response probabilities, P(rq | vj ). The virtual advisor obtains the prior probabilities for each individual from a logit model based on five truck characteristics: price, fuel economy, performance, reliability, and safety. Each customer is asked initial constant-sum, selfexplicated importance weights for these characteristics. (These prior weights are obtained from
questions that are asked prior to the question banks illustrated in Figure 3.) The prior probabilities are estimated with Equation 3 where the wc is the importance for the cth characteristic for
each individual obtained from a constant-sum allocation of 100 importance points across the five
scales, xjc is the value of characteristic c for vehicle vj, and β is a scaling parameter.
7
This reward function is related to the entropy function minimized by KW. However, the Bayesian advisor looks
two steps ahead rather than simply maximizing Equation 2 (minimizing entropy in KW). For earlier applications in
marketing of information-theory-based reward functions see Hauser (1978) and Herniter (1973). For recent applications in psychology, see Prelec (2001).
12
“Listening In” to find Unmet Customer Needs and Solutions
5
e
P (v j ) =
(3)
V
β ∑ wc x jc
c =1
5
∑e
β ∑ wc x jc
c =1
j =1
The characteristic values for each existing vehicle and the scaling parameters are obtained from
archival data and judgments by managers and engineers. For example, prior surveys to owners
help establish that the Toyota Tacoma 4x4 (regular cab) has a rating of 1.087 on fuel economy
and a rating of 1.241 on performance. For the GMC Sonoma 2WD Regular Cab the corresponding ratings are 2.116 and 0.525 respectively. (Data disguised slightly.) The actual data were synthesized from “an ongoing global effort” by the manufacturer “to understand consumers’ needs
and wants related to motor vehicles.” (Quotes from a proprietary study.) Part of this ongoing
global effort included data collected with the AIO questionnaire described earlier (76 “personal
viewpoints” and 38 “vehicle characteristics”). When new vehicles become available, managers
and engineers provide temporary estimates of the xjc’s.
The conditional response probabilities are also based on these ongoing AIO surveys, supplemented when necessary by experienced managers and engineers. For example, the survey
data suggest that customers who prefer the Toyota Tacoma 4x4 (regular cab) are likely to answer
that they prefer a four-wheel drive vehicle 84% of the time. They are likely to answer that they
prefer two-wheel drive only 16% of the time. Table 1 illustrates the type of data upon which the
conditional probabilities are based. These data are disguised slightly.
Table 1
Example Conditional Probabilities from AIO Survey, Supplemented with Judgment
Conditional Probability, P(rq | vj ) – data disguised
Number of
passengers
Chevy Avalanche 2WD
Chevy Silverado 2500 2WD
GMC Sonoma
4WD Crew Cab
(148 vehicles)
Dodge Ram
Club 4WD
1 passenger
5%
25%
15%
…
10%
2 passengers
15%
25%
5%
…
15%
3 passengers
25%
25%
15%
…
25%
4 passengers
25%
15%
25%
…
25%
5-6 passengers
30%
10%
25%
…
25%
13
…
“Listening In” to find Unmet Customer Needs and Solutions
Evolving Question Banks
Virtual advisors and the listening-in methods are not one-shot studies. Markets evolve as
customer needs change and technology improves. Each year brings changing characteristics and
new brands. To be effective in advising customers (virtual advisor) and identifying new opportunities (listening in), the virtual advisor must be relatively simple to update. For example, suppose that four-wheel steering becomes a feature that is important to customers (and a feature that
helps the advisor recommend a truck). Suppose further that some truck brands start offering this
feature for the 2003 model year. We can readily add a question bank on steering to the set of
available trucks. Because of the local independence property, we need only obtain incremental
data for the new question banks. We need to know how owners of each truck brand will rate
their vehicles on the new question bank. For new truck brands we need to know how owners of
the new brands will rate their vehicles on the characteristic values (xjc’s) and how they will answer each question bank. These data are obtained from the periodic AIO surveys – surveys that
are a normal part of business and used for many purposes. In essence, the virtual advisor and
listening in “free ride” on surveys undertaken by the manufacturer for other purposes.
Trigger Mechanism to Identify New Opportunities
Equation 2 enables the virtual advisor to select a question-bank order that leads to rapid
convergence toward recommendations. For many customers an existing vehicle will fulfill their
needs and the updated recommendation probabilities will evolve smoothly as in Figure 3. For
these customers we do not identify opportunities by listening in – they are satisfied and are not a
high-potential source of new ideas for product platforms. However, for some customers, their
answers to question banks reveal inconsistencies. For example, suppose that (1) the customer
has already answered constant-sum importance question banks that indicate reliability and low
price are most important (price 30 points, performance 10 points, fuel economy 20 points, reliability 30 points, and safety 10 points) and (2) the customer’s subsequent answers suggest an interest in a small truck with a 4-cylinder engine, two-wheel drive, and automatic transmission.
The Mazda B2300 fits these preferences best (see Figure 4 – Question banks 1 to 4). Given
these answers the virtual advisor decides that further information on towing and hauling will
clarify recommendations. The advisor expects that the customer will want to haul relatively light
loads such as small-garden equipment or tow a jet ski. Knowing the exact towing and hauling
needs will help the advisor decide among a number of otherwise comparable light-duty trucks.
14
“Listening In” to find Unmet Customer Needs and Solutions
However, suppose the customer says that he or she plans to use the truck to haul heavy
materials and tow a large motor boat (weighing 6,500 pounds). No existing light-duty truck can
tow such heavy loads effectively and safely. On the other hand, no truck that can tow such
heavy loads can fill the customer’s requirements as expressed in earlier question banks. This
may be an opportunity worth investigating – a light-duty truck that can occasionally haul heavy
materials or tow heavy loads.
The intuition in this example is that the question bank on towing and hauling revealed
something about the customer’s underlying needs. Based on this new information the customer
is probably not going to be satisfied with existing trucks and the virtual advisor will have to revise its best-truck recommendation probability downward. This drop in the maximum recommendation probability becomes a trigger for further investigation. We illustrate this trigger
mechanism by an arrow in the dialogue in Figure 4. Question Bank 5, which included questions
about towing and hauling, causes the most preferred vehicle to change from the Mazda to a Ford
Ranger (a slightly larger and more-powerful compact truck). The utility drops because this more
powerful compact truck cannot fully meet the towing and hauling requirements and because it
cannot meet the requirements expressed in Question banks 1-4. (It has a 6-cylinder engine and is
more expensive.) A full-sized truck, such as the Chevrolet Silverado 1500, could fulfill the towing and hauling requirements, but the advisor does not recommend the Silverado because it does
poorly on the other desired features. After further question banks the recommendation probabilities in Figure 4 increase because the Ford Ranger fulfills the additional requirements.
15
“Listening In” to find Unmet Customer Needs and Solutions
Figure 4
Example Use of the Trigger Mechanism
0.0533
Recommendation/Question Banks
Mazda B2300, Prior (Points *)
0.0735
Mazda B2300, Engine Size (4 cyl)
0.0861
Mazda B2300, Transmission (Auto, 2WD)
0.1105
Mazda B2300, Size (Compact)
0.1056
Ford Ranger, Towing/Hauling (Yes)
0.1200
Ford Ranger, Construction Plowing (No)
0.1243
Ford Ranger, Brand (All)
0.1328
Ford Ranger, Bed Length (Short)
0.1356
Ford Ranger, Tallest Person (< 6')
0.1401
Ford Ranger, Passengers (2)
0.1428
Ford Ranger, Maneuverability (Important)
0.1459
Ford Ranger, Big, Quiet (Neutral)
0.1478
Ford Ranger, Styling (Conventional)
0.1498
Ford Ranger, Price (20-22K)
0.00
0.04
0.08
0.12
0.16
Maximum Probability
The intuitive idea in Figure 4 has appeal, but before we incorporate the trigger mechanism we must investigate it further. For example, the posterior probability might drop because
there is error in the customer’s response. If the trigger mechanism is too sensitive, it might identify many false conflicts and the true conflicts might be lost in the noise. On the other hand, if it
is not sensitive enough, the trigger mechanism might miss opportunities. We show later in this
paper, through simulation, how to select a sensitivity level for the trigger mechanism such that
unmet-need segments are likely to be recovered. In these simulations we begin with real data for
the conditional probabilities and create known unmet-needs segments. We then add error and
examine how various sensitivity levels balance “false positives” and “false negatives.” The
simulations demonstrate that calibration is feasible and that the performance of the listening-in
mechanism is reasonably robust in the face of response errors. It is also reasonably robust with
respect to the sensitivity levels chosen for the trigger mechanism.
The other issue it theoretical. The intuition assumes that a drop in posterior probability
identifies a conflict in the underlying utility of the vehicle. If a question bank affected only the
vehicle that was recommended prior to the qth question bank and if that same vehicle were recommended after the qth question bank, then most random utility models would suggest that a
probability drop is an indicator of an underlying utility drop. For example, both the logit and the
16
“Listening In” to find Unmet Customer Needs and Solutions
probit models have this property. However, each question bank can affect the probabilities of all
148 vehicles and change the identity of the recommended vehicle based on the qth question bank.
We demonstrate formally in the Appendix that the intuition still holds. If the qth question bank
does not change the identity of the recommended vehicle, then a drop in posterior probability
indicates that the recommended vehicle has characteristics in conflict with the customer’s preferences. If the qth question bank changes the identity of the recommended vehicle, then a drop in
the posterior probability indicates that a truck with mixed characteristics would have higher utility than either the truck recommended before the qth question bank or the truck recommended
after the qth question bank. Both cases, if sustained across many customers, suggest opportunities for new products that satisfy unmet needs.
Identifying Potential Root Causes of the Utility Drops
When the trigger mechanism identifies a potential conflict, we need further information
to determine whether or not it is a true opportunity. We first identify which truck characteristics
are in conflict and then gather clarifying information from the customer.
As the Appendix establishes, conflicts exist when no existing truck simultaneously satisfies all of the customer’s needs. To diagnose such conflicts it is tempting to rely on “product archeology” to examine the correlations among the characteristics of trucks that are now on the
market (Ulrich and Pearson 1998). However, such “ecological” correlations represent more than
customer preferences; they represent the efficient frontier of the equilibrium responses by competing truck manufacturers. We prefer a mechanism that is less sensitive to supply-side and
equilibrium considerations. One such mechanism is the correlations in the underlying customer
preferences that drive the responses to the virtual advisor’s question bank. We obtain indicators
of these correlations by combining the virtual-advisor dialogue and the AIO data.
Let ρ rq rp be the correlation across vehicles of the conditional probabilities of a customer
answering rq to question bank q and answering rp to question bank p.8 Let Ρ be the matrix of
these correlations. For example, an element in this matrix might be the correlation across vehicles of the probabilities of a customer indicating that he or she (1) will use the truck for trailering
heavy loads and (2) prefers a rugged body style for that vehicle. Based on existing truck seg8
Such correlations across vehicles are consistent with local independence. The latter only assumes response independence conditioned on a given vehicle. Even with local independence we expect customers to be heterogeneous
and, hence, expect non-zero correlations across vehicles in customers’ answers to the question banks.
17
“Listening In” to find Unmet Customer Needs and Solutions
ments we expect these example characteristics to be positively correlated. On the other hand, a
priori, we expect a customer’s need to pull a large trailer or boat to be negatively correlated with
a preference for a compact truck. That is, we do not expect that customers who prefer compact
trucks will also value hauling and towing heavy loads. We expect their towing needs to be limited to lighter loads. (If the example motivating the drop in utility in Equation 3 occurs often in
the data, “listening in” will cause us to re-examine this a priori belief.) Thus, whenever the trigger mechanism suggests a potential opportunity, the listening-in algorithm examines all correlations corresponding to the customer’s answers to the first q question banks ( Rq−1 U rq ). It flags
those which are highly negative (less than –0.30 in our application). (The level of this flagging
mechanism is can also be set with simulation.) Such negative correlations trigger the virtual engineer.
A Virtual Engineer Clarifies the Opportunity
Customers who use a virtual advisor do so to make a better truck-purchasing decision;
they do not see themselves as respondents to a questionnaire. Thus, we must be careful to
choose our questions carefully. The listening-in trigger mechanism targets relatively-precisely
those customers who have unmet needs and the Ρ-triggering mechanism targets their unmet
needs. We now introduce a mechanism by which the listening-in methodology can concentrate
its questions to obtain relevant, more-detailed information about those unmet needs. We call this
mechanism a virtual engineer (VE). It asks relatively few questions of each targeted customer,
but, across many customers, its questions span the needs-space. The VE is also flexible; its
questions can be updated continuously without re-commissioning a large-scale AIO survey.
The concept of a VE is simple; its implementation difficult. To be useful to the productdevelopment team, the VE must ask the customer those questions that inform the engineering
design decisions that would be necessary should a promising opportunity be identified. To be
credible to the customer, the VE must ask questions in a non-technical manner that relates to
how the customer uses the truck. Naturally, the VE evolves through application, but we describe
here the process by which the initial VE questions are created.
For each potential conflict (negative ρ rq rp ), an engineering design team from a major
automotive manufacturer considered the basic engineering problem imposed by the conflicting
needs. The team then generated the questions that the team would need answered in order to de-
18
“Listening In” to find Unmet Customer Needs and Solutions
cide among basic solutions to that conflict. The engineering team formulated the questions that
they would ask the customer if they were participating in the dialogue between the advisor and
customer. For example, if the customer wants a compact truck that can tow a large boat, then the
engineering team would ask about the type of boat (e.g., modest sailboat, large motor boat, or
multiple jet skis) and the weight of the boat(s) that the customer plans to tow. The engineering
team would also ask the customer why he or she wants a compact truck (e.g., low price, tight
parking, high maneuverability, fuel economy, etc.). All engineering questions are then rephrased
into customer language.
Figure 5
Virtual Engineer
(a) Introductory Screen
(b) Example Dialogue
(c) Specific Questions to Elaborate
(d) Open-ended Questions
In addition to the questions identified by the engineering team, the VE includes openended dialogues which enable the customer to elaborate further the reasons underlying the previ19
“Listening In” to find Unmet Customer Needs and Solutions
ously-unidentified need. Figure 5 illustrates a sample dialogue in which the VE introduces himself, asks about a conflict, gathers quantitative data, and asks for open-ended comments. In this
example, the unmet-need conflict is between a full-sized truck and a 6-cyclinder engine.
A Design Palette Solicits Customer Solutions to Potential Conflicts
In our application, we supplemented the VE with a design palette. If the unmet customer
need is truly a promising opportunity, then it is important to explore that opportunity from multiple perspectives. One perspective is the customer’s own solutions – customers have proven to be
sources of new solutions in many categories including software, windsurfing, and mechanical
fasteners (von Hippel 1986; 1988; 2001a). Even for trucks, new ideas often come from user innovations at construction sites, trailering, camping, and racing (NASCAR’s Craftsman Truck
Series). Furthermore, such user solutions can be accelerated with innovation “tool-kits” which
enable customers to mix and match features (Franke and von Hippel 2002; von Hippel 2001b).
Such tool-kits are not unlike the mass-customization configurators used by Dell.com and Timbuk2.com in which customers pick and choose the features they want as they order their computer or luggage products. Such toolkits, known as user-design methods, have been used successfully in market research for instant cameras, laptop computer bags, crossover vehicles, ski resorts, high-speed copiers, and Internet Yellow Pages (Dahan and Hauser 2002; Leichty, Ramaswamy and Cohen 2001).
We illustrate one such “design palette” in Figure 6. Here the customer (a) receives instructions, (b) changes the size of the truck, and (c) changes the color. For brevity, we have not
shown the many intermediate steps, some of which include new state-of-the-art truck features
such as four-wheel steering and extra-wide frames. However, changes do not come free to the
customer. There are sophisticated engineering/cost models underlying the design palette. For
example, if the customer changes the size of the truck, then the price, fuel economy, and towing/payload capacity change accordingly. Once the customer completes the redesign he or she is
given the opportunity to indicate whether, and by how much, he or she prefers the new design.
(The customer may not prefer the new design because of accumulated “sticker shock” or because
of an holistic judgment of the final truck.) Nonetheless, in the empirical application described
20
“Listening In” to find Unmet Customer Needs and Solutions
later in this paper, 73% of the respondents who completed the exercise indicated that they would
purchase their custom-designed truck were it available.9
Figure 6
Design Palette
(a) Introductory Screen
(b) Customer Selects Size
(c) Customer Selects Color
(d) Customer Evaluates His or Her Design
Design palettes are evolving rapidly. For example, one system enables the customer to
adjust the length of the hood of a car or truck while the software automatically insures the integrity of other design elements such as the windshield angle and window shape. The customer simply clicks on the hood and drags it forward or clicks on the front bumper and pushes it back. Us-
9
Due to self-preference learning, memory accessibility, and context effects the preference for the self-designed
truck might be inflated (Bickart 1993; Feldman and Lynch 1988; Nowlis and Simonson 1997; Simmons, Bickart and
Lynch 1993; Tourangeau, Rips and Rasinski 2000). However, this does not diminish the value of the design palette
as a means to identify opportunities that will be explored in more depth when the product-development team cycles
through an iterative process.
21
“Listening In” to find Unmet Customer Needs and Solutions
ing this advanced design palette, the customer could create easily a Euro sports design (short
front overhang, high truck deck, low overall height) that is pleasing to the eye and incorporates
many “design” heuristics. Alternatively, by lengthening the front overhang and the hood the customer could create a classic look with a long sloping back to the truck. The software is sufficiently advanced that the customer could then rotate the model in all directions to get a full 3D
view.
Initial Sizing of the Opportunity
The virtual engineer and the design palette are triggered automatically whenever a probability drop is detected that is larger than a preset threshold. The virtual engineer is triggered for
(at most) six of the flagged conflict pairs to keep the respondent’s task relatively short. The limitation to six conflict pairs implies that no more than six input screens, such as Figure 5c, are presented to the customer. This limitation assures that the virtual engineer is not perceived as intrusive. The limitation was set based on pretests with customers.
While “listening in” can, in theory, identify all unmet-need combinations, not all such
combinations will justify further investigation. To make the decision on further investigation, a
truck manufacturer requires an initial estimate of the size of the opportunity. This estimate of potential can be a rough indicator because “listening in” is part of the fuzzy front end of an iterative
product development process. The manufacturer will evaluate any opportunities further before
any sizable investment. Fortunately, the listening-in methodology provides a method for initial
sizing that appears to be sufficient to distinguish the few big winners.
Subject to the caveats of self-selected customers and the approximations in Equations 13, we can identify patterns of unmet needs within the population. Each customer answers a custom-designed set of question banks. These question banks and the corresponding answers identify the customer’s needs. Let Ai represent customer i’s answers. Then Ai corresponds to a subset, Ρi, of the correlation matrix, Ρ. These data enable us to cluster respondents on Ρi to identify
groups of customers with similar needs. If the size of the unmet-need cluster is large, as a fraction of the initial sample, then this unmet-need segment is likely to be worth further investigation.10
10
“Listening in” is based on dialogues with virtual advisors. Because customers choose to initiate these dialogues,
there is self-selection. However, given the large fraction of truck customers who search for information on the web
22
“Listening In” to find Unmet Customer Needs and Solutions
Suppose we have identified an unmet-need segment. To simulate a truck “design” for
that segment we define a concept truck by the needs it fulfills. To estimate market share for the
concept truck we include the concept truck in the set of existing trucks available to the virtual
advisor. By using the Bayesian model in Equation 1, we calculate revised posterior probabilities
for all trucks, including the new truck concept. Averaging the revised posterior probabilities
over all respondents provides a rough estimate of market share for the new concept truck.
Monte Carlo Simulations: Sensitivity to Error and the Trigger Mechanism
We undertake Monte Carlo simulations to address three issues. The first set of simulations examines whether the listening-in methodology can recover unmet-need segments from
data provided by respondents who make errors in both the initial preference judgments (wc’s )
and the responses to the Bayesian advisor’s questions (rq’s ). The second set of simulations explores the sensitivity of the trigger mechanism and suggest that, over a reasonable range, this
trigger mechanism is robust with respect to the choice of trigger level. The third set of simulations vary the errors in the preference judgments and the question-bank responses to explore how
such errors affect performance. Together these simulations explore the internal validity of the
listening-in methodology and establish that it can identify promising opportunities in the presence of response error.
Simulation Methodology
We generated nine customer segments of 500 respondents each – a total of 4,500 simulated respondents. For six of the generated customer segments, respondents have preferences for
which no existing trucks satisfy their needs. If there were no error, customers in those segments
would answer the virtual advisor’s question banks according to their needs and the listening-in
methodology should correctly identify their unmet needs. For example in one segment we might
define an answer profile consistent with customers who want a compact truck that tows large
loads. If there were no response errors, customers in this segment would answer “yes” to “compact truck” and “yes” to the large-load question in the appropriate question bank. If no existing
truck meets both needs simultaneously, these answers would trigger a drop in the posterior probability and indicate unmet needs. In the simulations, we created segments to represent customers
who want (1) compact trucks that can haul and tow large loads, (2) sporty full-sized trucks with
and given the growth in virtual advisors, we expect the self-selection issues to diminish. At minimum, a large fraction of even the self-selected customers might still be an important opportunity for a new truck.
23
“Listening In” to find Unmet Customer Needs and Solutions
short beds, (3) compact trucks with diesel engines, (4) full-sized trucks with an extra-short bed
and 4-cylinder engines, (5) compact trucks with 10-cylinder engines, and (6) full-sized trucks
with high maneuverability. In each of these segments there were a number of true conflicts. For
example, simulated respondents in the first segment want a compact truck with a small engine
that can tow and haul large loads. At the “needs” level, this segment produces four conflict pairs
– compact truck/tow large loads, compact truck/haul large loads, 4-cylinder engine/tow large
loads, and a 4-cyclinder engine/haul large loads. For segments seven through nine, we created
profiles where existing trucks satisfy well the customers’ needs (e.g., full-sized trucks that can
tow and haul large loads). These three segments enable us to test whether or not the methodology identifies false opportunities.
For each respondent in the nine segment profiles we generated consistent responses, rq’s,
for each question bank and consistent self-explicated importances, wc’s. We then added errors to
the customer’s responses. There are two classes of question banks – question banks with nominal
categories and the question bank of constant-sum self-explicated importance questions. Because
the rq’s are nominal variables we assume that E% are answered incorrectly and that the incorrect
answers are uniformly distributed among the remaining categories. For example if the respondent truly wants a compact truck we simulate the a 10-percent error by having 10 percent of the
respondents answer that they want a large truck. Because the wc’s are interval-scaled variables
estimated by allocating 100 points across the five truck characteristics, we simulate response error in the these answers by adding a zero-mean, normally-distributed response error such that the
standard deviation of the error equals a specified number of points (e). For simplicity we truncate negative self-explicated importances which, fortunately, occur with low probability. We
then apply the listening-in equations to each of the 4,500 simulated respondents.
Whenever a probably drop occurs that is larger than the trigger level, we record the negative conditional-probability correlations, ρ rq rp , corresponding to all need-conflicts (unmet needs)
that are identified in the complete dialogue with that simulated respondent. If multiple trigger
points are identified, unmet needs for all triggers are recorded. Thus, each simulated respondent
who experiences one or more probability drops (above the trigger), is represented by a vector
which has negative ρ rq rp values for every pair of identified need conflicts. After simulating all
respondents, we cluster this conflict matrix to identify the unmet-need customer segments.
24
“Listening In” to find Unmet Customer Needs and Solutions
We use a k-means non-tree clustering algorithm based on the Euclidean norm defined on
the matrix of triggered correlations (respondents by potential conflict pairs). We use the standard “scree” rule to identify the number of clusters, n, but abstract n, n+1, and n+2 clusters. We
then examine their size and interpret their profiles. If the n+1st cluster is still large, we abstract
more clusters and continue until we do not find a large cluster. As a rule of thumb, to be sure
that no large unmet-need cluster is missed, we abstract additional clusters until the last two clusters are small and do not reflect interesting need patterns. This simulates the manner is which the
listening-in methodology is applied to actual data.
Initial Simulations to Uncover Unmet Needs
We begin with moderate error in both the self-explicated importances and the responses
to the question banks (E = 10% and e = 5 points). The VE is triggered and the conflict correlations are recorded whenever P ( v1 | rq , Rq−1 ) - P ( v1 | Rq−1 ) ≤ t where t = 0.00005. This is a rela-
tively sensitive trigger. We show later that this trigger is within the robust range.
Table 2 summarizes the results of the initial simulation for this trigger level. The entries
in this table indicate the number of respondents from a true segment that were assigned to a cluster. The largest number in each row is displayed in bold text. In total, 82.7% of the respondents
were correctly classified. Most of the misclassifications were respondents who had true unmet
needs (e.g., compact truck that tows large loads), but were classified to the null segment because
of errors in their responses. For example, a respondent might have had a true preference for a
compact truck that could tow large loads, but, due to errors, stated that he or she wanted a large
truck that could tow large loads. That respondent was assigned to the null segment because there
are many large trucks that can tow heavy loads. Such errors occur ten percent of the time in this
simulation. One perspective is the micro level and another is the macro level. First, at the micro
level, the simulation identified 21,096 conflict pairs when there were only 16,500 true conflict
pairs. (Recall that there are multiple conflict pairs for each of the 4,500 respondents.) This implies that 14% of the identified conflict pairs were false negatives and 36% were false positives.
Thus, at the micro level, the response errors imply significant errors in recovery. However, these
numbers do not tell the entire story. For managerial relevance we must look at the macro level.
25
“Listening In” to find Unmet Customer Needs and Solutions
Table 2
Results of the Simulated Cluster Analysis
Unmet-Need Segment
Cluster number:
Respondents Classified to Cluster
Total
1
2
3
4
5
6
7
8
9
Compact truck, large loads
418
0
0
1
0
0
81
0
0
500
Sporty full-sized, short bed
1
422
0
0
0
0
77
0
0
500
Compact truck, diesel
0
0
401
0
0
0
99
0
0
500
Full-sized, extra-short bed
1
0
0
346
0
0
153
0
0
500
Compact truck, 10 cylinders
3
27
0
0
336
0
134
0
0
500
Full-sized, maneuverable
0
2
0
0
0
346
92
43
17
500
Null segment
43
0
2
1
0
0
1454
0
0
1500
The managerial focus is to identify promising opportunities. To address this goal, we focus on each cluster and note those conflicts that were identified for the majority of the respondents in the cluster. For example, for the first cluster, the methodology identified the first unmetneed conflict (compact truck/tow large loads) for 95.9% of the respondents who had that conflict.
The other three conflicts in Segment 1 were identified 82.4%, 77.3%, and 73.3% of the time, respectively (review definitions earlier in this section). No other conflict was identified for more
than 9.4% of the respondents. Thus, for the first unmet-needs segment, despite errors in response, the listening-in methodology did extremely well on managerial recommendations – all
true conflict pairs and no false conflict pairs were identified by the majority of respondents in the
first cluster. This was true for all segments and corresponding clusters, including the null segment for which no conflict pairs were identified. The last two clusters would be considered too
small to represent any viable opportunities. Thus, at least for moderate levels of response errors
and for a sensitive trigger level, the listening-in methodology proved successful in terms of identifying unmet needs and unmet-need segments for further research.
Sensitivity to the Trigger Mechanism
In Table 3 we repeat the simulations for various trigger levels beginning with a trigger
that invokes the VE for any drop in posterior probabilities (t = 0.00000) to a trigger that is extremely insensitive to drops in posterior probabilities (t = 0.10000). The data suggest that the
trigger mechanism is best set to be sensitive. Indeed, classifications appear to be robust over a
moderate range of sensitive triggers. For example, in Table 2 the methodology was able to clas-
26
“Listening In” to find Unmet Customer Needs and Solutions
sify correctly 82.7% of the respondents. In Table 3 we obtain roughly the same percentage for
all triggers in the range of 0.00000 to 0.00100. More importantly, for this range of triggers we
identify correctly all of the unmet-need segments and all of the unmet-need opportunities. On
the other hand, if the trigger is not sufficiently sensitive we miss both segments and opportunities. We never identify false opportunities, even with an insensitive trigger, because the “majority-in-a-segment” criterion is extremely effective in terms of avoiding false positives.
Table 3
Calibrating the Trigger Mechanism
Trigger Level
Percent of
Respondents
Classified Correctly
Percent of
Opportunities
Identified Correctly
Percent of
Unmet-Needs
Segments Identified
False
Opportunities
Identified
t = 0.00000
82.73%
100%
100%
0
t = 0.00005
82.73%
100%
100%
0
t = 0.00010
82.69%
100%
100%
0
t = 0.00100
82.69%
100%
100%
0
t = 0.01000
56.69%
63.6%
63.4%
0
t = 0.10000
33.33%
0%
0%
0
Sensitivity to the Level of Response Errors
The third set of simulations examines the sensitivity of predictions to the levels of response error. Specifically, we simulate errors of 0%, 10%, and 20% for the nominal answers and
0 points, 5 points, and 10 points for the constant-sum answers. The results are summarized in
Table 4 for (a) the percent of respondents classified correctly and (b) the percent of unmet-needs
identified correctly. As expected, when there are no response errors, every customer is classified
correctly and all unmet needs are identified. Furthermore, classification and identification do not
appear to be particularly sensitive to errors in the self-explicated constant-sum responses (columns in Table 4). This is not surprising because the constant-sum responses are used for the
prior probabilities and, with enough question banks, the effect of prior probabilities diminishes.
Classification and identification are more sensitive to response errors in the virtual advisor’s question banks. If a respondent answers incorrectly 10% of the time, classification of responses are correct roughly 82% of the time. However, the algorithm still identifies correctly all
unmet-need opportunities.
27
“Listening In” to find Unmet Customer Needs and Solutions
The most severe condition in Table 4 is E=20%. In this condition respondents, on average, answer 1 in 5 questions incorrectly. Based on our observations and discussions with consumers, this error rate is much higher than we expect among real consumers who are actively
seeking information on which vehicle to purchase. For this error rate the classification rate drops
to 55-62%, but the methodology still identifies 75-94% of the unmet needs correctly.11
Table 4
Sensitivity to Response Errors
Errors in the Self-Explicated Importances (Priors)
Response Errors
(updating)
e = 0 points
e = 5 points
e = 10 points
E = 0%
100%
99.9%
99.9%
E = 10%
82.9%
82.7%
81.8%
E = 20%
61.6%
55.0%
56.8%
(a) Percent of respondents classified correctly
Errors in the Self-Explicated Importances (Priors)
Response Errors
(updating)
e = 0 points
e = 5 points
e = 10 points
E = 0%
100%
100%
100%
E = 10%
100%
100%
100%
E = 20%
93.9%
75.8%
81.8%
(b) Percent of needs correctly identified
Summary
Together Tables 2, 3, and 4 suggest a reasonable level of internal validity. In the presence of errors in both the prior preferences and the responses to the question banks, the listeningin algorithm appears to be able to identify promising opportunities for new truck platforms.
While the classification of respondents to segments is affected by these errors, unmet-need conflicts are still identified successfully. While recovery degrades for high levels of error in the
nominal responses, this level of recovery should be sufficient for the fuzzy front end of product
development. Because “listening-in” complements extant methods, final decisions on whether or
not to introduce new truck platforms can be tested with more extensive data collected later in the
11
There appears to be a slight anomaly in the last rows of Tables 4a and 4b. For E=20%, classification and identification appear to increase slightly with errors in the self-explicated importances. This happens because the combination of errors pushes more respondents to the “no-conflict” clusters. As a result, a few more “no-conflict” respon-
28
“Listening In” to find Unmet Customer Needs and Solutions
product-development process. The accuracy in Tables 2, 3, and 4 should be sufficient to focus
this complementary data collection on promising new “fishing grounds.”
Application to Identify New Opportunities for Pickup Truck Platforms
The initial application of “listening in” occurred in August of 2001 when 1092 customers
were recruited from the Harris Interactive Panel.12 The customers had purchased a pickup truck
in the past four years (1997-2000) and were given a $20 incentive for this initial test. Pickup
truck customers are a prime target for virtual advisors. For example, Scott Morton, Zettelmeyer,
and Silva-Risso (2001) report that the typical pickup buyer saves 2.9% of the purchase price with
an on-line service compared to an average of 1.5% for all vehicles.
On average each customer spent 45 minutes with the virtual advisor, design palette, and
virtual engineer (when triggered). Most customers found the experience worthwhile. Customers
trusted the virtual advisor by an 8-to-1 margin over dealers and would be more likely to purchase
a vehicle recommended by the virtual advisor by a 4-to-1 margin over a vehicle recommended
by a dealer. For the design palette, 78% found it an enjoyable experience and 82% felt it was a
serious exercise. When the virtual engineer was triggered, 88% found the questions easy to answer and 77% felt that the virtual engineer related well to their needs. Interestingly, 56% of the
customers reported that they would pay for the advise provided by the virtual advisor if it were
included in the price of the pickup truck that they purchased as a result of using the advisor.
In this initial application we set the trigger mechanism to be very sensitive to any utility
drop as suggested by Table 3. The most common pairwise conflicts were a maneuverable fullsized truck (38%), a compact truck which could tow and haul heavy materials (14%), and a fullsized truck with a six-cylinder engine (7%). Clustering pairwise conflict profiles, Ρi, identified a
segment of customers with unmet needs for large and maneuverable trucks. Of these respondents, ten percent wanted a top-of-the-line truck and sixteen percent wanted a standard full-sized
pickup. Another segment, thirteen percent of the respondents, had unmet needs for a compact
truck that could tow and haul heavy loads. Given the current engineering frontier, meeting these
needs would raise the price of the truck, thus not every respondent in the unmet-need segment
would purchase a new concept truck. We provide market share estimates below.
dents classified correctly. Two more unmet-needs are identified correctly because it is easier to achieve a “majority”
in the remaining clusters. Neither difference is significant at the 0.05 level with a two-tailed t-test.
12
This initial test was based on a stratified random sample of the panel. For this test, all customers were given the
opportunity to use the design palette.
29
“Listening In” to find Unmet Customer Needs and Solutions
As an illustration, Table 5 provides a perspective on the reasons that respondents cited
most when the virtual engineer sought further clarification of the full-sized-maneuverable-truck
unmet-need conflict. This qualitative input suggests that respondents are using the full-sized
truck for city driving. The large truck fulfills critical needs – they are willing to sacrifice maneuverability for large passenger capacity and large payloads. However, they would prefer maneuverability if they could get it.
Table 5
Elaboration of Customer Needs for a Full-Sized Maneuverable Pickup Truck
Why I need a maneuverable pickup truck.
Why I need a full-sized pickup truck.
Frequent city driving
66%
Large passenger capacity.
73%
Tight parking
58%
Large payloads.
50%
I make many U-turns
26%
Full-sized style.
39%
Too many traffic jams
28%
When the full-sized-maneuverable-truck segment of respondents were given the opportunity to redesign their most preferred pickup truck, the features that they changed most often were
truck height (6’ to 7’), truck width (6’ to 7’), and steering (two-wheel steering changed to fourwheel steering). This suggests that they are looking for an even larger truck, but that they would
be interested in four-wheel steering to gain maneuverability. Using the methods described earlier for market sizing we estimated the potential market share of a full-sized truck with fourwheel steering. Based on cost models, we calculated that the extra features would increase the
manufacturer’s suggested retail price (MSRP) by $3,000. For this concept truck, the listening-in
equations estimate a market-share increase for the manufacturer of 3-4% (the exact value is
coded for confidentiality).13 Such a $2.4-to-3.2 billion dollar per year opportunity is definitely
worth further investigation. In addition, a compact truck with heavy-duty hauling and towing is
estimated to be a $1-to-2 billion opportunity (values coded). Such unmet-need conflicts could be
fulfilled by a small truck platform with a strong frame, transmission, and engine.
Our initial application to pickup trucks illustrates that promising new-product opportuni13
We obtain rough forecasts by adding a full-sized maneuverable pickup truck to the choice sets of the unmet-needsegment customers. We obtain P(rq | vj) for the new vehicle by assuming a profile similar to an existing vehicle except for the critical responses on the size and maneuverability questions, which are changed to be consistent with the
vehicle being both full-sized and maneuverable. The iterative use of Equation 1 provides the estimates.
30
“Listening In” to find Unmet Customer Needs and Solutions
ties can be identified. After our study was complete we learned (previously unknown to us) that
the automotive manufacturer was in the process of introducing four-wheel steering in order to
improve maneuverability of its top-of-the-line pickup truck (the 2002 GMC Denali). This truck
is selling well. We plan to monitor the sales of this truck and, perhaps a basic full-sized truck
with four-wheel steering, to determine whether its sales are in the rough range predicted by the
market-sizing equations.
Summary, Discussion, and Future Research
In this paper we explore a methodology in which a virtual engineer “listens in” to a customer’s Internet dialogue with a trusted, virtual advisor. The use of such advisors is growing as
they become more effective, as the Internet itself gains further penetration, and as the value of
such advisors becomes recognized by customers. The fraction of people using the Internet for
information and advice is large (62% in autos, 70% in travel, and 56% in health) and advisors are
becoming more common. “Listening in” provides a means to capture the information in these
dialogues – data that are readily available with little incremental cost to the researcher. It is
likely that the listening-in methodology can be applied successfully for these and other complex
customer decisions.
There are five modules to the listening-in methodology – the Bayesian advisor, the listening-in trigger, the virtual engineer, the design palette, and clustering of identified unmet-need
conflicts. This combination of methods, tested in Monte Carlo simulations and applied in a
“proof-of-concept” demonstration, appears useful for automotive applications. In other applications some modules can be replaced. For example, the Bayesian advisor might be replaced with
conjoint-analysis-based advisors. Alternatively, sub-matrix clustering of the unmet-need conflicts might be replaced with latent-segment analyses.
Like all methodologies, “listening in” will benefit from continuous improvement. The
initial application builds on existing methods that are used in new ways. Each component can be
improved – better methods to identify priors, more efficient look-ahead algorithms, improved
calibration of the trigger mechanism, and better indicators of conflicting needs could all benefit
from further research. The dialogues, the user interfaces, and the presentation of stimuli are all
areas of potential improvement. For example, work is now underway to put more “stretch” into
the design palette and to give the virtual advisors and the virtual engineers personalities based on
“talking heads.”
31
“Listening In” to find Unmet Customer Needs and Solutions, References
References
Balasubramanian, Siva K. and Wagner A. Kamakura (1989), “Measuring Consumer Attitudes Toward the
Marketplace with Tailored Interviews,” Journal of Marketing Research, 26, (August), 311-326.
Barabba, Vincent P. (2003), Anticipate and Lead: The Sustained Competitive Advantage of KnowledgeBased Adaptive Enterprise, forthcoming.
---- and Gerald Zaltman (1991), Hearing the Voice of the Market: Competitive Advantage Through Creative Use of Information, (Boston MA: Harvard Business School Press).
Bickart, Barbara A. (1993), “Carryover and Backfire Effects in Marketing Research,” Journal of Marketing Research, 30, 1, (February), 52-62.
Blackorby, C., D. Primont, and R. R. Russell (1975), “Budgeting, Decentralization, and Aggregation,”
Annals of Economic Social Measurement, 4, 1, 23-24.
Burchill, Gary W. (1992), Concept Engineering: The Key to Operationally Defining Your Customers’
Requirements. (Cambridge, MA: Center for Quality Management).
Crawford, C. Merle (1991), New Products Management, Third Edition, (Homewood, IL: Richard D. Irwin, Inc.), 123-129.
Dahan, Ely and John R. Hauser (2002), “The Virtual Customer,” Journal of Product Innovation Management, 19, 5, (September), 332-353.
Dolan, Robert J. (1993), Managing the New Product Development Process, (Reading, MA: AddisonWesley Publishing Co.).
Feldman, Jack M. and John G. Lynch, Jr. (1988), “Self-generated Validity: Effects of Measurement on
Belief, Attitude, Intention, and Behavior,” Journal of Applied Psychology, 73, (August), 421-435.
Franke, Nikolaus and Eric von Hippel (2002), “Satisfying Heterogeneous User Needs via Innovation
Toolkits: The Case of Apache Security Software,” MIT Sloan School of Management Working
Paper #4341-02, Cambridge, MA 02421.
Gallagher, Robert (1968), Information Theory and Reliable Communication, (New York, NY: John Wiley
& Sons).
Green, Paul E. and V. Srinivasan (1990), “Conjoint Analysis in Marketing: New Developments With Implications for Research and Practice,” Journal of Marketing, pp. 3-19.
Griffin, Abbie J. and John R. Hauser (1993), “The Voice of the Customer,” Marketing Science, Winter,
pp. 1-27.
Gutman, Jonathan (1982), “A Means-End Chain Model Based on Customer Categorization Processes.”
Journal of Marketing, 46, (Spring), 60-72.
Hauser, John R. (1978), "Testing the Accuracy, Usefulness and Significance of Probabilistic Models: An
Information Theoretic Approach," Operations Research, Vol. 26, No. 3, (May-June), 406-421.
R1
“Listening In” to find Unmet Customer Needs and Solutions, References
______ and Glen L. Urban (1986), "Value Priority Hypotheses for Consumer Budget Plans," Journal of
Consumer Research, Vol. 12, No. 4, (March), 446-462.
Herniter, Jerry (1973), “An Entropy Model of Brand Purchase Behavior,” Journal of Marketing Research,
10, 361-375.
J. D. Power and Associates (2001), “New Autoshopper.com Study,” (Agoura Hills, CA: J. D. Power and
Associates).
Kamakura, Wagner A. and Michel Wedel (1995), “Life-Style Segmentation with Tailored Interviewing,”
Journal of Marketing Research, 32, (August), 308-317.
Kolpasky, Kevin (2002), “NPD Practices: Bringing Ideas to the Development Stage at General Motors,”
PDMA Visions, 26, 1, (January), 13-15.
Kullback, S. (1954). Information Theory and Statistics, (New York: John Wiley & Sons).
Lehmann, Donald R. and Russell S. Winer (1994), Product Management, (Boston, MA: Richard D. Irwin, Inc.).
Liechty, John, Venkatram Ramaswamy, Steven Cohen (2001), “Choice-Menus for Mass Customization:
An Experimental Approach for Analyzing Customer Demand With an Application to a Webbased Information Service,” Journal of Marketing Research, 38, 2, (May).
McFadden, Daniel (1974), “Conditional Logit Analysis of Qualitative Choice Behavior,” Frontiers in
Econometrics, P. Zarembka, ed., (New York: Academic Press), 105-142.
Narasimhan, Chakravarthi and Subrata K. Sen (1983), “New Product Models for Test Market Data,”
Journal of Marketing47, 1, (Winter), 11-24.
Nowlis, Stephen M. and Itamar Simonson (1997), “Attribute-Task Compatibility as a Determinant of
Consumer Preference Reversals,” Journal of Marketing Research, 36, (May), 205-218.
Plummer, Joseph T. (1974), “The Concept and Application of Life Style Segmentation,” Journal of Marketing, 38, (January), 33-37.
Prelec, Drazen (2001), “A Two-person Scoring Rule for Subjective Reports,” Working Paper (Cambridge, MA: MIT Sloan School).
Savage, L. J. (1971). “Elicitation of Personal Probabilities and Expectations,” Journal of the American
Statistical Association, 66, 783-801.
Scott Morton, Fiona, Florian Zettelmeyer, and Jorge Silva-Risso (2001), “Internet Car Retailing,” Journal
of Industrial Economics, 49, 4, 501-519.
Shocker, Allan D. and V. Srinivasan (1979), “Multiattribute Approaches to Product Concept Evaluation
and Generation: A Critical Review,” Journal of Marketing Research, 16, (May), 159-180.
Simmons, Carolyn J., Barbara A. Bickart, and John G. Lynch, Jr. (1993), “Capturing and Creating Public
Opinion in Survey Research,” Journal of Consumer Research, 30, (September), 316-329.
R2
“Listening In” to find Unmet Customer Needs and Solutions, References
Singh, Jagdip, Roy D. Howell, and Gary K. Rhoads (1990), “Adaptive Designs for Likert-Type Data: An
Approach for Implementing Marketing Surveys,” Journal of Marketing Research, 27, (August)
304-321.
Srinivasan, V. (1982), “Comments on the Role of Price in Individual Utility Judgments,” in Choice Models for Buyer Behavior, Leigh McAlister, ed., (Greenwich, CT: JAI Press, Inc.), 81-90.
Time (1938), "Henry Grady Weaver - 2,000,000 Opinions Make a Fact", Time Magazine, 32, 20, (Nov. 14)
Tourangeau, Roger, Lance J. Rips, and Kenneth A. Rasinski (2000), The Psychology of Survey Response,
(New York, NY: Cambridge University Press).
Ulrich, Karl T. and Scott Pearson (1998), “Assessing the Importance of Design Through Product Archaeology,” Management Science, 44,3, (March),352-369.
Urban, Glen L. and John R. Hauser, Design and Marketing of New Products, 2E, (Englewood Cliffs, NJ:
Prentice Hall, Inc.).
______, ______, and John. H. Roberts (1990), "Prelaunch Forecasting of New Automobiles: Models and
Implementation," Management Science, Vol. 36, No. 4, (April), 401-421.
______, Fareena Sultan, and William J. Qualls (2000), “Placing Trust at the Center of Your Internet
Strategy,” Sloan Management Review, 42, 1, (Fall), 39-48.
______, Bruce Weinberg, and John R. Hauser (1996), "Premarket Forecasting of Really-New Products,"
Journal of Marketing, 60,1, (January), 47-60. Abstracted in the Journal of Financial Abstracts, 2,
23A, (June) 1995.
von Hippel, Eric (1986), “Lead Users: A Source of Novel Product Concepts,” Management Science, 32,
791-805.
______ (1988), The Sources of Innovation, (New York, NY: Oxford University Press).
______ (2001a), “Innovation by User Communities: Learning from Open-Source Software,” MIT Sloan
Management Review, 42, 4, (Summer), 82-86.
______ (2001b), “Perspective: User Toolkits for Innovation,” Journal of Product Innovation Management, 18, 247-257.
Wind, Jerry (1982), Product Policy, (Reading, MA: Addison-Wesley, Inc.).
Zaltman, Gerald (1997), “Rethinking Market Research: Putting People Back In,” Journal of Marketing
Research, 23 (November), 424-437.
R3
“Listening In” to find Unmet Customer Needs and Solutions, Appendix
Appendix: Formal Motivation of Trigger Mechanism
The “listening-in” methodology uses a Bayesian trigger mechanism in which the virtual
engineer and design palette are triggered whenever the posterior recommendation probability
(Equation 1) drops. We argue intuitively in the text that such a drop in the recommendation
probability is an indication of an opportunity for improving a product. In this appendix we demonstrate with a formal analytical model that such a drop identifies at least some opportunities.
This issue is not trivial because a question bank, q, affects, potentially, the updated utilities of
each and every product in the market, not just the recommended product.
Although our application uses complex question banks for 148 trucks, we can illustrate
the basic principles with N = 3 and a dichotomous question bank. (Our propositions generalize
to analogs for larger N and for polychotomous question banks, but the notation is cumbersome.)
Following the text, let j index the vehicles. Without loss of generality, let v1 be the recommended
r
product after question bank q-1. Let x j be those truck characteristics that are not affected by
r
question bank q and let y j be those truck characteristics that are affected by question bank q. In
r
r
this formulation, price is treated as a characteristic and can be in either x j or y j (for motivation
see Srinivasan (1982) and Hauser and Urban (1996). Following Blackorby, Primont and Russell
r r
r
r
(1975) we model preferences by a utility tree such that u( x j , y j ) = ux( x j ) + uy( y j ) + ε, where ε
is a Gumbel-distributed error term that represents the uncertainty in utility due to question banks
that have not yet been asked (or may never be asked). For simplicity we assume that trucks with
r
r
r
r
y j = y good experience an increase in utility and trucks with y j = ybad experience a decrease in
utility. (The dichotomous question bank reveals which trucks have desirable characteristics.)
We let v2 be a surrogate for those products with desirable characteristics and v3 be a surrogate for
those products with undesirable characteristics (as revealed by question bank q). Based on
McFadden (1974) we write the recommendation probabilities in more-fundamental utility-theory
terms.
(A1)
P ( v j | rq , Rq−1 ) =
e
V
r
r
ux ( x j )+u y ( y j )
∑e
r
r
u x ( xm ) + u y ( y m )
=
e
r
r
u x ( x1 ) + u y ( y1 )
e
+e
r
r
ux ( x j )+u y ( y j )
r
r
u x ( x2 ) + u y ( y good )
+e
r
r
u x ( x3 ) +u y ( ybad )
m =1
After question bank q two situations can occur. Either the recommended truck re-
A1
“Listening In” to find Unmet Customer Needs and Solutions, Appendix
r r
mains v1 or the recommended truck becomes v2 . It cannot become v3 because, even if y1 = ybad ,
v1 would still be preferred to v3 .
Proposition 1. If the recommended truck after question bank q is the same truck as that
r
recommended after question bank q-1, then v1 has undesirable characteristics ( y1 =
r
ybad ) if and only if P ( v1 | rq , Rq−1 ) decreases.
Proposition 2. If the recommended truck after question bank q is different than the truck
recommended after question bank q-1 and if the recommendation probability der r
creases, then (1) v1 has undesirable characteristics ( y1 = ybad ) and (2) a truck with mixed
r
r
characteristics, x1 and y 2 , would have higher utility than the both the recommended truck
after q-1 question banks and the recommended truck after q question banks.
Proofs. Straightforward algebra establishes that P ( v1 | rq , Rq−1 ) - P ( v1 | Rq−1 ) is proportional to
e
r
r
r
u x ( x1 ) +u x ( x3 ) + u y ( y good )
−e
r
r
r
u x ( x1 ) + u x ( x3 ) +u y ( ybad )
r
r
r
r
r
r
r r
u ( x ) +u ( x ) + u ( y )
u ( x ) + u ( x ) +u ( y
)
≥ 0 if y1 = y good and e x 1 x 2 y bad - e x 1 x 2 y good ≤
r r
0 if y1 = ybad . Algebra also establishes that the proportionality (denominator) is positive. This
establishes Proposition 1. If the recommended truck changes after question bank q, then
P ( v1 | rq , Rq−1 ) < P( v2 | rq , Rq−1 ) and because the recommendation probability decreases we have
P( v2 | rq , Rq−1 ) < P ( v1 | Rq−1 ) . Thus P ( v1 | rq , Rq−1 ) < P ( v1 | Rq−1 ) and, by Proposition 1, we have
r r
that y1 = ybad . This establishes the first result in Proposition 2. Because v1 was recommended
r
r
r
r
before question bank q, we have u x ( x1 ) > u x ( x2 ) and by supposition we have u y ( y good ) - u y ( ybad )
r
r
> 0, thus a product with the features, x1 and y 2 , will have higher utility than either v1 or v2 . This
establishes the second result in Proposition 2.
Generalizations. If there are n2 trucks like v2 and n3 trucks like v3, then the analogs to
Propositions 1 and 2 are readily proven. The numbers n2 and n3 enter the equations for
P ( v1 | rq , Rq−1 ) - P ( v1 | Rq−1 ) , but the basic proofs remain intact. If there are many trucks
r
r
r
with y good or ybad , but with different x j , the expressions for P ( v1 | rq , Rq−1 ) - P ( v1 | Rq−1 ) include more terms, but each can be proven to have the correct sign – increases if
r
r
y good and decreases if ybad . With these changes, the remaining portions of the proofs
follow as above.
A2