The Way Entrepreneurs Learn: Migration of a 30
Transcription
The Way Entrepreneurs Learn: Migration of a 30
The Way Entrepreneurs Learn: Migration of a 30-Second Elevator Pitch Exercise from Instructor-Led Delivery to Blended Learning Technology Dan Suhr, (inXsol, LLC), Sharon Ballard (EnableVentures, Inc.), Ricardo Gomez (NCIIA), Henry Ryng (inXsol, LLC), Angela Shartrand (NCIIA) Abstract inXsol was awarded an NSF SBIR Grant (0945987)1 to research migration of established entrepreneur education (EE) tools, Supercoach® Entrepreneurial Training2 (SET), from an instructor-led to a blended learning environment. One part of the project was converting a 30Second Elevator Pitch exercise to an online module that familiarized students with sample pitches, clarified the key components of an elevator pitch and allowed the students to video and share their own pitch through the module. In addition, an innovative assessment protocol was designed so that the student 30-Second Elevator Pitches were electronically reviewed by a panel of industry and EE experts using a rubric specifically designed for this project. This paper will identifies why this exercise was selected, the key learning objects, user experience for the module, the assessment strategies embraced and the commercialization lessons learned. Introduction The objective for the National Science Foundation (NSF) Small Business Innovation Research (SBIR) Phase I entrepreneur education (EE) project was to develop scaleable, compelling, interdisciplinary curricula and delivery methods that provide broader access to curriculum and reduce the entrepreneur’s time from "innovative idea to market implementation." To achieve those goals inXsol, with Henry Ryng serving at the project’s Principle Investigator, proposed to determine the efficacy of web-based simulations and exercises with the Supercoach® Entrepreneurial Training (SET) curriculum that has historically been delivered in an instructorled format. The proposed research addressed the question: Can migrating a portion of the SET curriculum from instructor-led delivery to web-based delivery shorten the duration of courses and workshops and lead to significantly improved EE throughput while retaining (or increasing) the participant’s performance levels? Body The development of the 30-Second Elevator Pitch was one component of the SBIR project. The product resulting from the Phase I effort was a web-based treatment of two SET exercises (30Second Elevator Pitch and the One-Page Strategic Business Plan) traditionally presented by a qualified instructor in classes and workshops. We called the resulting product SET Academy 1 NSF GRANT SUPPORT AND DISCLAIMER -‐ The project described was supported by Grant Number 0945987 from the National Science Foundation. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation. 2 Supercoach® Entrepreneurial Training was developed by Sharon Ballard of EnableVentures, Inc. and licensed to inXsol, LLC for NSF SBIR Grant (0945987) development. (www.setacademy.com). The 30-Second Elevator Pitch exercise was developed with Level 2 interaction characteristics while the One-Page Strategic Business Plan EE exercise was developed with Level 1 interaction characteristics. Levels of Interactivity are general classes of richness, sophistication and realism of interactivity the participant experiences in eLearning and simulations. Level 1 has very limited interactions while Level 4 would be a free play simulation where the participant is not limited by predetermined paths and there are high levels of interaction. After the modules were developed the assessment protocol was designed compare a control cohort (workshop 1) with a test cohort (workshop 2). The test cohort was asked to complete the two web-based exercises before attending a SET workshop. The control cohort was not exposed to these web-based exercises before attending the SET workshop. Rather, the control cohort was given the same pre-workshop materials traditionally provided. The two 4-day workshops were conducted about a month apart. Both workshops used the same instructor, facilities, guest speakers and “dream panel” of live reviewers. The educational model depicted below is a progression structure where students are exposed to the EE material and exercises prior to the class or workshop, proceed to live instructor-led workshops and classes, move on to a series of one-on-one coaching sessions and then are finally in their own. For this project we envisioned EE exercises and simulations leading the student through the process with a collaborative thread and cumulative exercises that would support the student through the progression. inXsol’s experience is that this supported structure helps normalize student knowledge and allows for the instructor to address the most complex information instead of using valuable time to review foundational information. Figure 1 – Proposed Structure inXsol’s experience supports this approach. During the last 15 years inXsol has developed simulations with Level 1 to 4 interactions mimicking the performance and operation of aviation and industrial machinery. Simulations are useful when training on “live” equipment is too dangerous, expensive or scarce to use for operator training. These simulations were structured to embrace experiential learning concepts advanced by Kolb and Fry (1975). The experiential learning model they described included four elements: concrete experience, observation and reflection, the formation of abstract concepts and testing in new situations (Kolb and Fry 1975). This model, although somewhat simplistic, describes the rationale behind inXsol’s hard skill simulators developed for our customers. inXsol’s experience with complex simulations is easily transferrable to the experiential learning required for EE. inXsol uses scaffolding strategies (Bransford et al. 2000) to support a student while learning separate components of new materials in our simulations and exercises. The characteristics of scaffolding that advanced this project included: • • • Model progression. Learning environments using model progression are a major design framework for both classroom and eLearning applications. A learning objective is broken into chunks and presented with more and more content or complexity included as student performance and mastery progresses. Use of domain specific content. Access to domain-specific information is provided through questions and answers, reference materials, web-links, and inspirational entrepreneurial stories. Adult professionals gather information on their own, and the mastery of the literature of a profession is a key skill for long-term success. Engaging assignments. The learning environment provided engaging assignments using authentic scenarios that provided the situations, questions and exercise situations used to motivate the user in a variety of situations. These assignments are embedded in the learning environments with model progression as noted above. Finally, inXsol’s experience shows that significant efficiencies can be gained for both the facilitator and the student by moving proven instructor-led SET workshop to an interactive online infrastructure. However, it is not just inXsol that have found these efficiencies in eLearning and simulations. Setaro (2001) summarized the studies that examined the training compression experienced by companies that adopted web-based delivery of training. Fletcher (1990) after carefully reviewing over forty independent studies found that Technology Based Training (TBT) yielded time savings of 35-45% over traditional classroom instruction while obtaining equivalent or better gains in learning retention and transfer. Adams (1992), Cantwell (1993) Bradley (1994) and Hofstetter (1994) later confirmed these results across several years and hundreds of studies. Likewise Hall’s (1997) in-depth review of over 130 case studies found that computer-based training required significantly less time than instructor-led training. The amount of reduction in time ranged from 20-80% with 40-60% being the most common range for timesaving. None of the studies and meta-studies reported a decrease in training effectiveness and most reported a substantial increases in training effectiveness (both learning retention and transfer). More recently, Hemphill (1997) found that while CBT saves time, it does not negatively impact effectiveness of learning. (Setaro 2001) These characteristics and inXsol’s experience in developing training solutions, coupled with the SET experience with entrepreneurial training and development, provided compelling reasons for the NSF to award the proposal. The purpose of our research was to answer the following sets of research questions that supported our primary question. First, we posed questions in context of the test cohort as compared to the control cohort. Those questions included: • How can web-based exercises/simulations be incorporated in the experiential learning of SET? • Was a participant’s performance in the workshop enhanced by the web-based exercises completed prior to their participating in the instructor-led course or workshop? • Did the individuals in the test cohort begin the workshop with consistent understanding of the SET process and expectations of the workshop as measured through comparison of the workshop pretests? The theory was that the test cohorts’ workshop pretest scores would be clustered with less variance in performance than the control cohorts. • Would a test cohort perform measurably better than the control cohort? • What were the differences in course evaluations supplied by both the test cohort and control cohort? Our second set of research questions addressed the test cohort’s experience with the Level 1 exercise interaction characteristics compared to the exercise developed to Level 2 interaction standards: • How much time did each test subject spend completing or interacting with each exercise? • What was the pre-test and post-test performance delta for the differing levels of interaction? • Did the interactivity level of the web-based exercises result in different observable inworkshop performance? • Using test cohort survey results, were participants more likely to complete a web-based exercise if there is Level 2 interactions present compared to Level 1 interactions? In summary, we expected to determine how web-based exercises and simulations could enhance the learning process in entrepreneurs. We also established how these tools and technologies could supplement the experiential learning environment in traditional SET training. Our focus was on developing qualitative metrics for demonstrating the efficacy of our tools. Module Development Two exercises were selected to convert to web-based experiential learning modules: 30-Second Elevator Pitch and the One-Page Strategic Business Plan. The criteria used to select these exercises out of dozens of potential exercises included: sequence in the workshop (early exercises in the workshop would better isolate the impact of the online modules), interactivity (interactive exercises provide an opportunity to better demonstrate the effectiveness of the experiential simulation approach), and “hooks” (exercises that had components that would be used in later workshop exercises would provide a foundation for additional future exercises and also allow for the participant’s constituency to be evaluated). Rubrics were designed to evaluate each exercise. While is it important to describe the criteria used to select the modules for the NSF SBIR project, the remainder of the paper will focus on one exercise – the 30-Second Elevator Pitch. 30-Second Elevator Pitch Module. This exercise was chosen because it is the first step in the Supercoach® process as illustrated below and compared to the standard business planning process. The Supercoach® process relies on oral communication skills and builds towards writing a business plan when one is needed for financing or strategic alliances; whereas, the traditional process the written business plan is the first task undertaken - arguably the most difficult task of the process. The Supercoach® process allows an idea to be focused through short exercises. When a student completed the 30-Second Elevator Pitch module, they know the components of an elevator pitch, will have evaluated at least nine pitches, and, finally, recorded their own pitch. The 30-Second Elevator Pitch was divided into three rounds of interactions. The first round the student watches nine videos of actual pitches of startup companies. From those nine pitches they would select five to promote to the next round where they will better understand evaluation criteria. Below are screen captures from round one and round two. Figure 2 – 30-Second Elevator Pitch Example of Module Round two includes more detailed evaluation of the five pitches that were promoted using the following criteria. Criteria Market Opportunity? Was the value proposition of their opportunity clear? What is the problem they are solving? Was the size of their opportunity clear and quantified? Was the target customer of their opportunity identified? Why you? Was their competitive advantage clear? Did they express why their management team is qualified for their opportunity? Impact? Did they explicitly request some form of support (time, expertise or money)? Were they memorable? Could you repeat the main ideas of their venture? Was it 30-seconds or less? Good Story? Did their pitch arouse your attention and hold it? Would you share their story with others? Compelling? Was their argument persuasive and forceful? Did you learn something? After the student scores each video, the coach’s perspective becomes available. That perspective is an evaluation of the video using the coaching philosophy of asking questions and not lecturing or advising. This process help the student understand and evaluate the criteria. Finally, it is the student’s turn to create and practice their pitch in the third round of the module. Five different scenarios were created to put the student in a situation where they may be called upon to deliver their elevator pitch. The system check the user’s computer for a microphone and web camera and then provides a way for the student to record their pitch and play the pitch back at a later point. It was planned that in subsequent versions of the SET Academy the user will be able to allow their team members, coaches or instructors to view and provide feedback on the recording. The following screenshot illustrates the interface for round three of the 30-Second Elevator Pitch Module. The coach’s perspective in this round provides a recap of the evaluation criteria for a 30-Secnd Elevator Pitch. It was anticipated that this module would take 30-45 minutes for a student to complete. Figure 3 – 30-Second Elevator Pitch – “It’s Your Turn” Evaluation Design for the 30-SecondElevator Pitch Module The evaluation utilized a non-equivalent post-test design (quasi-experimental). Recruitment and assignment of workshop attendees was managed by inXsol with the help of Arizona State University. Workshop Participants Attendees were randomly selected to participate in the workshops through recruitment efforts of Arizona State University (ASU) SkySong. ASU advertised the workshop opportunity on their website and through their network of advisors that serve the greater Phoenix entrepreneurial community. The only limitation inXsol placed on recruiting was that no more than half of the workshop participants would be student teams. Attendees were assigned on a rolling basis to workshop 1 (comparison group), unless that did not work with the attendees’ schedule, in which case they were assigned to workshop 2 (treatment group). The recruitment for workshop 2 was undertaken to reflect the demographic mix attained in the first workshop in terms of student vs. community participants. The demographics of the two cohorts were similar except for two characteristics: education and work experience. In the case of education, in workshop 1 (comparison) 56 percent of the group had completed some graduate work or had an advanced degree. For workshop 2, 83 percent of the participants had undertaken graduate work or completed an advanced degree. The other characteristic that stood out was Entrepreneurship Experience. Forty-four percent of workshop 1 participants reported holding leadership roles in a start-up, while workshop 2 only reported 33 percent of the participants holding a leadership role in a start-up. Finally, 31 percent of the workshop 1 participants reported exiting a start-up enterprise through sale or merger compared to 17 percent for workshop 2 participants. Generally speaking, it appears workshop 1 had more start-up experience and workshop 2 reported greater education levels. Scoring of Student Samples: Online and Live Judging Two simple instruments were used to measure the quality of work samples. One five-item checklist was created to score the 30-Second Elevator Pitch. Twenty-seven judges were recruited to score the quality of work samples via a web-based tool. Three of the judges scored both types of work samples (elevator pitch and five-minute presentation), and the other 24 were randomly assigned to score either the pitch or presentation exercise. Each judge was given a random selection of work samples from each exercise that they were asked to score using the above checklist or rubric as appropriate. Two types of work samples were judged: a) a video of 30Second Elevator Pitch and b) a five-minute video of a final presentation). To minimize procedural confounds, work samples were created and presented in an identical manner (e.g., video clips used the same backdrop, audio, and video settings). Judges logged into the website and were given instructions to score the work samples. Judges were instructed to view the elevator pitch or final presentation video, and then were presented with a scoring sheet to enter their scores for each. When one work sample was scored, the judge had the option to continue scoring another work sample, or to quit the session. The number of work samples judges chose to score varied. Judges scored between three and 24 of the 30-Second Elevator Pitches) and between three and 19 of the five-minute presentations. [Below are screen captures of the online evaluation tool created for this NSF Phase I project.] Figure 4 – 30-Second Elevator Pitch – Evaluator Interface Major Findings Resulting from these Activities The panel of 15 judges independently assessed the 30-second pitches, which were delivered by 24 participants. In total, judges submitted 358 scores (224 for the comparison and 134 for the treatment) using the five-item checklist (see Figure 2). Each item had three categories of performance “Yes”(11-5), “No” (0-5), and “Needs Work” (6-10). Individual scores were calculated for each of the items, and the final score was made up of the scores in the five items. The final scores are the basic unit of analysis (n=358). The result of the independent t-test shows no significant statistical difference between the treatment (M=10.54, SD=2.53) and comparison group (M=10.16, SD=2.37); t(356)=---‐1.44, p=0.214 (95%CI ---‐0.90 to 0.14). Summary of Research In summary, no differences were found between judges’ ratings on the elevator pitches delivered by workshop participants in each group. The treatment participants performed as well as the comparison group verifying the research hypotheses. From a practical standpoint that means that over 2 hours of the instructor-led course was replaced by 30-45 minute eLearning module that the students completed before the workshop began. This freed the instructor to expose the students to more complex subjects or for more one-on-one instruction. As a result, students that took the online modules prior to attending the workshop exhibited higher performance in the subsequent exercises including the 5-Minute final presentation. There may be several explanations for this finding. In the elevator pitch, judges may focus on public speaking abilities and the quality of oral communication skills. These skills may not be readily improved in a short time period, especially when there is a 30 second time limit. Also, we did not account for differences in participants’ backgrounds that may have influenced their oral communications skills prior to the workshops. For example, we did not ask participants about how much prior experience they had with public speaking, nor did we ask about their English language abilities or experience developing elevator pitches in the past. These prior experiences are likely to be a strong influence on performance in the exercise and should be controlled for in future studies. Other suggestions for future work included ensuring that participants would be provided with the rubric and checklist beforehand so they can prepare their presentations based on the criteria upon which they are going to be evaluated later on. It would be worthwhile also to ensure that when online participants do the online exercises, they have the opportunity to apply the same rubric to the presentations or pitches they are rating. Conclusion Entrepreneurial Education has been identified as a key building block in the entrepreneurial process for scientists and engineers. inXsol has shown that technology can be used to develop scaleable, compelling, interdisciplinary curricula and delivery methods providing broader access to the target audience and compression of the entrepreneurial process duration from "innovative idea to market implementation." By making the process more efficient, the community gains through the jobs created by these new institutions. If the best ideas can be advanced more rapidly through the process, then job creation and economic growth will accelerate. Better access to innovative training will allow this to happen. There are three components of the EE process that can be enhanced providing better outcomes for the country’s entrepreneurs: curriculum, delivery, and collaboration. In the past the Supercoach® curriculum and its innovative approach to support the scientist and engineer entrepreneur has been proven a more effective training method than traditional approaches. inXsol’s Phase I research demonstrated that engaging, interactive web-based modules can compress the training time or delivery for the student and increase instructors’ capacity. inXsol believes that future research will show that the efficiency and quality can be extracted from the collaboration process – medium-term support for the students or venture teams through ongoing interaction with coaches. Unfortunately, NSF did not fund the collaboration approach outlined and submitted by inXsol for the Phase II proposal. The most important conclusion for the Phase I research is the interest the marketplace has exhibited for a technology-assisted approach to entrepreneurial education – especially more highly interactive approaches. One example of this is the enthusiasm in which many people were enthusiastically participating in this project – the two workshops (18 participants), live evaluators (4), electronic evaluators (27) and a four-hour focus group with the project board of advisors. However, the interest never converted into commercial subscriptions to the modules. Of the four commercial enterprises that expressed interest in subscribing to these prototype modules during the Phase I project only one organization actually paid the modest fee to expose their students to the modules. How could a product with so much promise fail in the EE marketplace? There seemed to be three primary reasons the EE organizations hesitated subscribing to the modules. First, there was an element of the “not invented here” mentality. Potential customers liked what they saw but wanted to tweak the modules one way or another without even a modest budget for redevelopment. Speaking of money, the institutions most interested in the modules were hesitant to pay a subscription even for the modest pricing of the modules. Those institutions did not want to pay, during a transition period, for the traditional paper-based SET curriculum and additional electronic modules to support the traditional SET curriculum. Which ties into the final barrier to adoption – availability of all the modules. In all instances, the institutions wanted the total curriculum available online as a condition of subscription purchases. Therefore, once the NSF Phase II funding was not available to complete the curriculum, the institutions were unwilling to fund the remaining development. Highly interactive eLearning modules and simulations are the way of the future. The industry is moving that direction but there is still the necessity to demonstrate an ROI for the customer and also overcome the threat EE instructors feel about a portion of their instruction replaced by eLearning modules. The 30-Second Elevator Pitch module was innovative and effective when it was created and the modules are still recognized for innovation and instructional design creativity. References Adams, G. L. 1992. "Why interactive?" Multimedia & Videodisc Monitor. (March). Bradley Associates. 1994. Multimedia made easy: Guide to developing interactive multimedia for training. Palo Alto, CA: Bradley Associates. Bransford, J. D., Brown, A. L., & Cocking, R. R. 2000. How People Learn: Brain, Mind, Experience, and School. Expanded Edition. Cantwell, S. 1993. "Multimedia transforms Union Pacific's training strategy." Tech Trends. (November/December). Fletcher, J.D. 1990. Effectiveness and cost of interactive videodisc instruction in defense training and education. Washington DC: Institute for Defense Analyses. (July). Hall, B. 1995. "Multimedia training's return on investment." Workforce Training News. (July/August). Hemphill, H. H. 1997. The impact of training on job performance. Scottsdale, AZ: Thompson NETg. Hofstetter, F. T. 1994. "Is multimedia the next literacy?" Educator's Tech Exchange. Inside Technology Training. Louisville, KY: Ziff-Davis. (Winter). Kolb. D. A. and Fry, R. 1975. 'Toward an applied theory of experiential learning”, in C. Cooper (ed.) Theories of Group Process, London: John Wiley. Kuratko, Donald F. 2005. “The Emergence of Entrepreneurship Education: Development, Trends, and Challenges” Entrepreneurship Theory & Practice, Vol 29 #5: 577-598. Katz, J.A. 2003. “The Chronology and Intellectual Trajectory of American Entrepreneurship Education.” Journal of Business Venturing, 18, 283-300 Kuratko, Donald. 2003. “Entrepreneurship Education: Emerging Trends and Challenges for the 21st Century.” Coleman Foundation White Paper Series for the U.S. Association of Small Business & Entrepreneurship in cooperation with The Entrepreneurship Program College of Business Ball State University. Mache, D. and D. Markley. 2004. “A County Guide: Energizing an Entrepreneurial Economy,” Lincoln, NE, Center for Rural Entrepreneurship. Raffo, C., A. Lovatt, et al. 2000. "Teaching and learning entrepreneurship for micro and small businesses in the cultural industries sector," Education and Training (6): 356 – 365] Setaro, John. 2001. Many Happy Returns: Calculating E-Learning ROI. [online]: http://www.learningcircuits.org/2001/jun2001/Elearn.htm [referred: 27.09.2005] Stacy, W., Walwanis Jacobs, M., Colanna-Romano, J. 2007. “Using pedagogical information to provide more effective scenarios.” Proceedings of the Interservice/Industry Training, Simulation & education Conference (I/ITSEC). Orlando, FL. Yaw, D. C. (2005). An evaluation of e-learning in industry at Level Three based upon the Kirkpatrick model. (Doctoral dissertation, Indiana State University, 2005). Dissertation Abstract International, A66/12, 136.